[ 461.482395] env[62558]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62558) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 461.482745] env[62558]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62558) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 461.482874] env[62558]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62558) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 461.483220] env[62558]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 461.578719] env[62558]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62558) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 461.588824] env[62558]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62558) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 462.189946] env[62558]: INFO nova.virt.driver [None req-cc239126-bc78-479f-983a-600a127c68ed None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 462.260649] env[62558]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 462.260892] env[62558]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 462.260968] env[62558]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62558) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 465.398714] env[62558]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-e6a67789-bbc3-4858-82ad-629b94bcb648 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.414363] env[62558]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62558) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 465.414548] env[62558]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b277b02b-f43c-4e2d-bbaa-c59faee8ca51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.446952] env[62558]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 7d6d4. [ 465.447176] env[62558]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.186s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 465.447605] env[62558]: INFO nova.virt.vmwareapi.driver [None req-cc239126-bc78-479f-983a-600a127c68ed None None] VMware vCenter version: 7.0.3 [ 465.450900] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca9bfaf-218c-4cc0-9f3e-a746dcc0db91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.472013] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0064429-4150-4410-8b32-5b51f815d3fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.477923] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f57c093-be25-42b4-86c0-590edb5b22df {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.484437] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3d34c7-698a-42b2-90da-82a7b396345d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.497193] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855503d1-af99-4f24-a587-a1fbb3206d6f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.502878] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a9543e-6f13-4e69-af37-42d668cab61a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.531753] env[62558]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-a32dd2b7-523c-46c6-b626-9afb838d8bc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 465.536524] env[62558]: DEBUG nova.virt.vmwareapi.driver [None req-cc239126-bc78-479f-983a-600a127c68ed None None] Extension org.openstack.compute already exists. {{(pid=62558) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 465.539161] env[62558]: INFO nova.compute.provider_config [None req-cc239126-bc78-479f-983a-600a127c68ed None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 466.042541] env[62558]: DEBUG nova.context [None req-cc239126-bc78-479f-983a-600a127c68ed None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),d4ed48a8-9e90-495c-9862-306001a7bafe(cell1) {{(pid=62558) load_cells /opt/stack/nova/nova/context.py:464}} [ 466.044614] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.044833] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.045547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 466.045976] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Acquiring lock "d4ed48a8-9e90-495c-9862-306001a7bafe" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.046181] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Lock "d4ed48a8-9e90-495c-9862-306001a7bafe" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.047178] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Lock "d4ed48a8-9e90-495c-9862-306001a7bafe" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 466.067146] env[62558]: INFO dbcounter [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Registered counter for database nova_cell0 [ 466.075072] env[62558]: INFO dbcounter [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Registered counter for database nova_cell1 [ 466.078495] env[62558]: DEBUG oslo_db.sqlalchemy.engines [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62558) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 466.078848] env[62558]: DEBUG oslo_db.sqlalchemy.engines [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62558) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 466.083681] env[62558]: ERROR nova.db.main.api [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 466.083681] env[62558]: result = function(*args, **kwargs) [ 466.083681] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 466.083681] env[62558]: return func(*args, **kwargs) [ 466.083681] env[62558]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 466.083681] env[62558]: result = fn(*args, **kwargs) [ 466.083681] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 466.083681] env[62558]: return f(*args, **kwargs) [ 466.083681] env[62558]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 466.083681] env[62558]: return db.service_get_minimum_version(context, binaries) [ 466.083681] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 466.083681] env[62558]: _check_db_access() [ 466.083681] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 466.083681] env[62558]: stacktrace = ''.join(traceback.format_stack()) [ 466.083681] env[62558]: [ 466.084428] env[62558]: ERROR nova.db.main.api [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 466.084428] env[62558]: result = function(*args, **kwargs) [ 466.084428] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 466.084428] env[62558]: return func(*args, **kwargs) [ 466.084428] env[62558]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 466.084428] env[62558]: result = fn(*args, **kwargs) [ 466.084428] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 466.084428] env[62558]: return f(*args, **kwargs) [ 466.084428] env[62558]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 466.084428] env[62558]: return db.service_get_minimum_version(context, binaries) [ 466.084428] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 466.084428] env[62558]: _check_db_access() [ 466.084428] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 466.084428] env[62558]: stacktrace = ''.join(traceback.format_stack()) [ 466.084428] env[62558]: [ 466.084853] env[62558]: WARNING nova.objects.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Failed to get minimum service version for cell d4ed48a8-9e90-495c-9862-306001a7bafe [ 466.084969] env[62558]: WARNING nova.objects.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 466.085408] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Acquiring lock "singleton_lock" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 466.085569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Acquired lock "singleton_lock" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 466.085808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Releasing lock "singleton_lock" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 466.086141] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Full set of CONF: {{(pid=62558) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 466.086284] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ******************************************************************************** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 466.086407] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Configuration options gathered from: {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 466.086541] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 466.086732] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 466.086857] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ================================================================================ {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 466.087077] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] allow_resize_to_same_host = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087247] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] arq_binding_timeout = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087373] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] backdoor_port = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087495] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] backdoor_socket = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087650] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] block_device_allocate_retries = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087806] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] block_device_allocate_retries_interval = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.087969] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cert = self.pem {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088146] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088320] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute_monitors = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088478] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] config_dir = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088666] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] config_drive_format = iso9660 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088809] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.088976] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] config_source = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089156] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] console_host = devstack {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089318] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] control_exchange = nova {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089474] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cpu_allocation_ratio = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089630] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] daemon = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089791] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] debug = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.089943] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_access_ip_network_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090114] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_availability_zone = nova {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090264] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_ephemeral_format = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090415] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_green_pool_size = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090640] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090799] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] default_schedule_zone = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.090954] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] disk_allocation_ratio = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091132] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] enable_new_services = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091316] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] enabled_apis = ['osapi_compute'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091474] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] enabled_ssl_apis = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091634] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] flat_injected = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091790] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] force_config_drive = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.091947] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] force_raw_images = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.092134] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] graceful_shutdown_timeout = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.092289] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] heal_instance_info_cache_interval = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.092501] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] host = cpu-1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.092675] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.092837] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] initial_disk_allocation_ratio = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093034] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] initial_ram_allocation_ratio = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093251] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093427] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_build_timeout = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093587] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_delete_interval = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093752] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_format = [instance: %(uuid)s] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.093949] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_name_template = instance-%08x {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094131] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_usage_audit = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094302] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_usage_audit_period = month {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094467] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094627] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] instances_path = /opt/stack/data/nova/instances {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094794] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] internal_service_availability_zone = internal {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.094990] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] key = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095122] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] live_migration_retry_count = 30 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095290] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_color = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095449] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_config_append = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095613] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095770] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_dir = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.095925] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096063] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_options = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096239] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_rotate_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096419] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_rotate_interval_type = days {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096583] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] log_rotation_type = none {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096709] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096832] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.096998] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097177] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097305] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097470] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] long_rpc_timeout = 1800 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097629] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_concurrent_builds = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097784] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_concurrent_live_migrations = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.097940] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_concurrent_snapshots = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098110] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_local_block_devices = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098266] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_logfile_count = 30 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098419] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] max_logfile_size_mb = 200 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098575] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] maximum_instance_delete_attempts = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098740] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metadata_listen = 0.0.0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.098903] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metadata_listen_port = 8775 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099082] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metadata_workers = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099246] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] migrate_max_retries = -1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099412] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] mkisofs_cmd = genisoimage {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099616] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] my_block_storage_ip = 10.180.1.21 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099749] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] my_ip = 10.180.1.21 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.099936] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] network_allocate_retries = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100101] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100275] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] osapi_compute_listen = 0.0.0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100432] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] osapi_compute_listen_port = 8774 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100603] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] osapi_compute_unique_server_name_scope = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100792] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] osapi_compute_workers = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.100959] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] password_length = 12 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101132] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] periodic_enable = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101290] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] periodic_fuzzy_delay = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101456] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] pointer_model = usbtablet {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101624] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] preallocate_images = none {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101781] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] publish_errors = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.101909] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] pybasedir = /opt/stack/nova {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102076] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ram_allocation_ratio = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102238] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rate_limit_burst = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102402] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rate_limit_except_level = CRITICAL {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102558] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rate_limit_interval = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102713] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reboot_timeout = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.102866] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reclaim_instance_interval = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103031] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] record = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103201] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reimage_timeout_per_gb = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103360] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] report_interval = 120 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103518] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rescue_timeout = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103674] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reserved_host_cpus = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.103830] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reserved_host_disk_mb = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104015] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reserved_host_memory_mb = 512 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104180] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] reserved_huge_pages = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104338] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] resize_confirm_window = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104496] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] resize_fs_using_block_device = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104648] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] resume_guests_state_on_host_boot = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104813] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.104982] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] rpc_response_timeout = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105157] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] run_external_periodic_tasks = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105324] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] running_deleted_instance_action = reap {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105482] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] running_deleted_instance_poll_interval = 1800 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105637] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] running_deleted_instance_timeout = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105792] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler_instance_sync_interval = 120 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.105954] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_down_time = 720 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106133] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] servicegroup_driver = db {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106309] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] shell_completion = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106471] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] shelved_offload_time = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106627] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] shelved_poll_interval = 3600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106793] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] shutdown_timeout = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.106950] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] source_is_ipv6 = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.107123] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ssl_only = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.107371] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.107537] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] sync_power_state_interval = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.107695] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] sync_power_state_pool_size = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.107862] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] syslog_log_facility = LOG_USER {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108029] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] tempdir = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108193] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] timeout_nbd = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108359] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] transport_url = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108517] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] update_resources_interval = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108674] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_cow_images = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108829] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_eventlog = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.108982] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_journal = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109151] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_json = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109307] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_rootwrap_daemon = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109460] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_stderr = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109612] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] use_syslog = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109762] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vcpu_pin_set = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.109924] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plugging_is_fatal = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110098] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plugging_timeout = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110259] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] virt_mkfs = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110414] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] volume_usage_poll_interval = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110569] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] watch_log_file = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110734] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] web = /usr/share/spice-html5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 466.110926] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.111106] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.111270] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.111437] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_concurrency.disable_process_locking = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.111988] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.112195] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.112368] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.112545] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.112740] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.112921] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.113121] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.auth_strategy = keystone {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.113294] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.compute_link_prefix = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.113471] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.113645] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.dhcp_domain = novalocal {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.113811] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.enable_instance_password = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114014] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.glance_link_prefix = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114203] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114377] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114586] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.instance_list_per_project_cells = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114703] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.list_records_by_skipping_down_cells = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.114905] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.local_metadata_per_cell = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115120] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.max_limit = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115296] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.metadata_cache_expiration = 15 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115473] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.neutron_default_tenant_id = default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115645] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.response_validation = warn {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115815] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.use_neutron_default_nets = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.115982] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.116158] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.116359] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.116541] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.116712] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_dynamic_targets = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.116875] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_jsonfile_path = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117070] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117266] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.backend = dogpile.cache.memcached {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117434] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.backend_argument = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117605] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.config_prefix = cache.oslo {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117773] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.dead_timeout = 60.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.117936] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.debug_cache_backend = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118111] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.enable_retry_client = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118272] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.enable_socket_keepalive = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118446] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.enabled = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118599] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.enforce_fips_mode = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118763] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.expiration_time = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.118924] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.hashclient_retry_attempts = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119100] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.hashclient_retry_delay = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119264] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_dead_retry = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119420] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_password = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119581] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119740] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.119899] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_pool_maxsize = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120068] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120231] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_sasl_enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120407] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120569] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_socket_timeout = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120728] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.memcache_username = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.120892] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.proxies = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121065] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_db = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121224] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_password = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121391] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_sentinel_service_name = mymaster {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121562] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121730] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_server = localhost:6379 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.121923] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_socket_timeout = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122103] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.redis_username = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122266] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.retry_attempts = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122428] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.retry_delay = 0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122590] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.socket_keepalive_count = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122749] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.socket_keepalive_idle = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.122908] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.socket_keepalive_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123076] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.tls_allowed_ciphers = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123234] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.tls_cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123389] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.tls_certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123546] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.tls_enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123699] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cache.tls_keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.123878] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124070] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.auth_type = password {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124237] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124419] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.catalog_info = volumev3::publicURL {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124573] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124755] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.124931] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.cross_az_attach = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125107] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.debug = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125268] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.endpoint_template = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125429] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.http_retries = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125587] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125741] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.125911] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.os_region_name = RegionOne {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126086] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126246] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cinder.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126416] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126573] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.cpu_dedicated_set = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126727] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.cpu_shared_set = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.126891] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.image_type_exclude_list = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127064] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127226] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.max_concurrent_disk_ops = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127387] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.max_disk_devices_to_attach = -1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127547] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127717] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.127879] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.resource_provider_association_refresh = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128051] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128216] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.shutdown_retry_interval = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128397] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128573] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] conductor.workers = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128749] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] console.allowed_origins = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.128910] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] console.ssl_ciphers = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129088] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] console.ssl_minimum_version = default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129258] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] consoleauth.enforce_session_timeout = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129422] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] consoleauth.token_ttl = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129594] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129750] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.129911] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130077] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130236] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130392] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130548] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130700] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.130858] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131026] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131176] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131330] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131482] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131646] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.service_type = accelerator {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131823] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.131995] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.132198] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.132406] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.132596] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.132758] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] cyborg.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.132936] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.backend = sqlalchemy {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133118] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.connection = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133284] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.connection_debug = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133449] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.connection_parameters = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133604] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.connection_recycle_time = 3600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133762] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.connection_trace = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.133952] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.db_inc_retry_interval = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134135] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.db_max_retries = 20 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134298] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.db_max_retry_interval = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134470] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.db_retry_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134638] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.max_overflow = 50 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134788] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.max_pool_size = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.134948] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.max_retries = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135128] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135284] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.mysql_wsrep_sync_wait = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135439] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.pool_timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135597] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.retry_interval = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135750] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.slave_connection = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.135907] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.sqlite_synchronous = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136077] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] database.use_db_reconnect = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136274] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.backend = sqlalchemy {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136452] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.connection = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136616] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.connection_debug = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136782] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.connection_parameters = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.136943] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.connection_recycle_time = 3600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137117] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.connection_trace = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137279] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.db_inc_retry_interval = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137439] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.db_max_retries = 20 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137598] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.db_max_retry_interval = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137757] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.db_retry_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.137913] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.max_overflow = 50 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138085] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.max_pool_size = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138247] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.max_retries = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138412] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138568] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138723] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.pool_timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.138883] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.retry_interval = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139050] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.slave_connection = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139212] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] api_database.sqlite_synchronous = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139386] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] devices.enabled_mdev_types = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139561] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139731] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ephemeral_storage_encryption.default_format = luks {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.139892] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ephemeral_storage_encryption.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140063] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140233] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.api_servers = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140397] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140555] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140716] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.140874] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141038] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141203] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.debug = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141365] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.default_trusted_certificate_ids = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141524] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.enable_certificate_validation = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141681] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.enable_rbd_download = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141835] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.141998] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142176] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142330] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142484] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142644] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.num_retries = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142808] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.rbd_ceph_conf = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.142966] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.rbd_connect_timeout = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143148] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.rbd_pool = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143315] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.rbd_user = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143475] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143630] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143783] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.143977] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.service_type = image {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.144164] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.144331] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.144488] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.144644] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.144823] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.145021] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.verify_glance_signatures = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.145186] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] glance.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.145397] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] guestfs.debug = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.145571] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] mks.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.145975] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.146249] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.manager_interval = 2400 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.146448] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.precache_concurrency = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.146622] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.remove_unused_base_images = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.146792] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.146959] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147153] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] image_cache.subdirectory_name = _base {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147327] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.api_max_retries = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147491] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.api_retry_interval = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147648] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147808] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.auth_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.147964] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148136] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148299] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148461] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.conductor_group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148617] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148772] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.148930] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149100] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149258] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149413] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149566] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149729] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.peer_list = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.149884] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150051] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150217] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.serial_console_state_timeout = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150373] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150538] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.service_type = baremetal {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150695] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.shard = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.150854] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151022] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151177] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151332] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151509] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151668] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ironic.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.151849] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152026] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] key_manager.fixed_key = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152219] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152388] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.barbican_api_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152544] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.barbican_endpoint = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152710] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.barbican_endpoint_type = public {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.152865] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.barbican_region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153030] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153188] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153346] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153500] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153667] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153806] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.number_of_retries = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.153992] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.retry_delay = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.154177] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.send_service_user_token = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.154335] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.154490] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.154646] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.verify_ssl = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.154799] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican.verify_ssl_path = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155024] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155132] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.auth_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155289] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155446] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155647] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155813] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.155968] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156146] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156328] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] barbican_service_user.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156497] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.approle_role_id = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156655] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.approle_secret_id = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156822] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.kv_mountpoint = secret {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.156983] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.kv_path = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157162] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.kv_version = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157321] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.namespace = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157478] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.root_token_id = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157634] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.ssl_ca_crt_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157797] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.timeout = 60.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.157957] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.use_ssl = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.158171] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.158378] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.158549] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.auth_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.158709] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.158865] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159040] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159238] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159402] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159559] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159716] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.159870] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160035] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160194] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160349] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160503] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160656] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160822] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.service_type = identity {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.160981] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161152] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161308] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161463] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161641] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161795] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] keystone.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.161994] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.connection_uri = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162168] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_mode = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162332] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_model_extra_flags = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162497] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_models = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162664] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_power_governor_high = performance {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162830] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_power_governor_low = powersave {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.162988] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_power_management = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.163171] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.163336] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.device_detach_attempts = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.163495] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.device_detach_timeout = 20 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.163658] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.disk_cachemodes = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.163816] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.disk_prefix = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164038] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.enabled_perf_events = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164226] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.file_backed_memory = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164394] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.gid_maps = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164553] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.hw_disk_discard = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164709] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.hw_machine_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.164879] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_rbd_ceph_conf = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165060] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165226] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165393] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_rbd_glance_store_name = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165561] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_rbd_pool = rbd {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165729] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_type = default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.165886] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.images_volume_group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166058] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.inject_key = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166258] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.inject_partition = -2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166482] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.inject_password = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166657] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.iscsi_iface = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166820] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.iser_use_multipath = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.166985] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_bandwidth = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167165] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167330] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_downtime = 500 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167491] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167649] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167806] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_inbound_addr = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.167966] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.168147] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_permit_post_copy = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.168307] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_scheme = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.168500] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_timeout_action = abort {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.168683] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_tunnelled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.168846] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_uri = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.169014] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.live_migration_with_native_tls = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.169180] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.max_queues = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.169342] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.169584] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.169748] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.nfs_mount_options = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170045] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170227] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170424] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_iser_scan_tries = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170590] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_memory_encrypted_guests = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170754] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.170916] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_pcie_ports = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.171097] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.num_volume_scan_tries = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.171267] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.pmem_namespaces = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.171425] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.quobyte_client_cfg = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.171711] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.171886] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rbd_connect_timeout = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172064] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172232] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172391] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rbd_secret_uuid = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172547] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rbd_user = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172705] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.172874] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.remote_filesystem_transport = ssh {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173045] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rescue_image_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173206] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rescue_kernel_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173362] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rescue_ramdisk_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173527] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173681] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.rx_queue_size = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.173861] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.smbfs_mount_options = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.174156] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.174332] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.snapshot_compression = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.174492] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.snapshot_image_format = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.174708] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.174873] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.sparse_logical_volumes = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175047] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.swtpm_enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175275] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.swtpm_group = tss {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175429] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.swtpm_user = tss {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175598] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.sysinfo_serial = unique {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175758] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.tb_cache_size = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.175915] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.tx_queue_size = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176095] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.uid_maps = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176279] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.use_virtio_for_bridges = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176459] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.virt_type = kvm {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176627] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.volume_clear = zero {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176788] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.volume_clear_size = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.176951] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.volume_use_multipath = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.177124] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_cache_path = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.177291] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.177455] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_mount_group = qemu {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.177619] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_mount_opts = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.177784] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178072] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178276] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.vzstorage_mount_user = stack {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178469] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178649] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178823] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.auth_type = password {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.178981] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179151] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179312] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179467] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179620] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179785] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.default_floating_pool = public {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.179941] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180119] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.extension_sync_interval = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180280] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.http_retries = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180441] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180596] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180748] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.180914] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181080] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181247] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.ovs_bridge = br-int {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181411] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.physnets = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181579] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.region_name = RegionOne {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181737] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.181902] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.service_metadata_proxy = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182071] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182239] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.service_type = network {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182424] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182591] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182747] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.182903] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183094] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183256] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] neutron.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183427] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] notifications.bdms_in_notifications = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183599] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] notifications.default_level = INFO {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183766] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] notifications.notification_format = unversioned {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.183962] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] notifications.notify_on_state_change = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.184149] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.184325] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] pci.alias = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.184494] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] pci.device_spec = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.184662] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] pci.report_in_placement = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.184867] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185056] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.auth_type = password {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185228] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185386] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185541] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185698] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.185853] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186014] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186183] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.default_domain_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186358] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.default_domain_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186516] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.domain_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186669] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.domain_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186821] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.186977] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187143] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187295] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187444] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187607] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.password = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187760] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.project_domain_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.187923] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.project_domain_name = Default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188101] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.project_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188274] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.project_name = service {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188441] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.region_name = RegionOne {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188599] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188755] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.188922] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.service_type = placement {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189093] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189252] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189406] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189559] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.system_scope = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189710] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.189862] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.trust_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190022] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.user_domain_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190191] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.user_domain_name = Default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190380] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.user_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190566] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.username = nova {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190746] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.190908] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] placement.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191095] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.cores = 20 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191261] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.count_usage_from_placement = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191427] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191601] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.injected_file_content_bytes = 10240 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191766] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.injected_file_path_length = 255 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.191934] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.injected_files = 5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192114] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.instances = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192282] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.key_pairs = 100 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192444] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.metadata_items = 128 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192606] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.ram = 51200 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192763] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.recheck_quota = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.192927] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.server_group_members = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193104] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] quota.server_groups = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193285] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193445] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193630] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.image_metadata_prefilter = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193779] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.193964] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.max_attempts = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.194152] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.max_placement_results = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.194318] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.194503] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.query_placement_for_image_type_support = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.194666] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.194837] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] scheduler.workers = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195017] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195195] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195371] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195537] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195697] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.195859] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196030] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196243] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196427] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.host_subset_size = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196593] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196751] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.196911] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197090] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.isolated_hosts = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197263] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.isolated_images = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197425] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197582] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197743] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.197902] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.pci_in_placement = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198074] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198237] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198398] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198556] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198714] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.198874] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199043] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.track_instance_changes = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199222] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199392] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metrics.required = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199555] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metrics.weight_multiplier = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199715] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.199877] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] metrics.weight_setting = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.200204] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.200379] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.200556] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.port_range = 10000:20000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.200726] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.200894] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201072] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] serial_console.serialproxy_port = 6083 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201238] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201408] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.auth_type = password {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201566] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201721] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.201883] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202052] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202213] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202414] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.send_service_user_token = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202592] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202767] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] service_user.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.202938] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.agent_enabled = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.203114] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.203422] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.203615] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.203784] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.html5proxy_port = 6082 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.203968] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.image_compression = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204175] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.jpeg_compression = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204308] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.playback_compression = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204469] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.require_secure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204635] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.server_listen = 127.0.0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204803] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.204962] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.streaming_mode = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205138] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] spice.zlib_compression = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205304] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] upgrade_levels.baseapi = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205471] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] upgrade_levels.compute = auto {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205628] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] upgrade_levels.conductor = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205782] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] upgrade_levels.scheduler = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.205944] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206116] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.auth_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206305] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206474] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206635] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206794] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.206950] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207125] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207283] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vendordata_dynamic_auth.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207453] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.api_retry_count = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207610] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.ca_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207776] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.cache_prefix = devstack-image-cache {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.207940] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.cluster_name = testcl1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.208115] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.connection_pool_size = 10 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.208272] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.console_delay_seconds = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.208489] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.datastore_regex = ^datastore.* {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.208721] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.208899] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.host_password = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209078] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.host_port = 443 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209250] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.host_username = administrator@vsphere.local {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209416] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.insecure = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209577] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.integration_bridge = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209737] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.maximum_objects = 100 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.209892] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.pbm_default_policy = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210061] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.pbm_enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210220] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.pbm_wsdl_location = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210384] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210537] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.serial_port_proxy_uri = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210689] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.serial_port_service_uri = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.210850] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.task_poll_interval = 0.5 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211022] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.use_linked_clone = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211192] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.vnc_keymap = en-us {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211353] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.vnc_port = 5900 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211511] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vmware.vnc_port_total = 10000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211692] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.auth_schemes = ['none'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.211873] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.212189] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.212358] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.212525] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.novncproxy_port = 6080 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.212699] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.server_listen = 127.0.0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.212868] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213038] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.vencrypt_ca_certs = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213200] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.vencrypt_client_cert = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213375] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vnc.vencrypt_client_key = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213533] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213694] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_deep_image_inspection = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.213885] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214049] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214220] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214400] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.disable_rootwrap = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214580] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.enable_numa_live_migration = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214741] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.214899] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215072] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215233] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.libvirt_disable_apic = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215390] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215548] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215707] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.215867] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216035] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216198] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216355] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216512] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216669] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.216832] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217019] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217188] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.client_socket_timeout = 900 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217353] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.default_pool_size = 1000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217515] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.keep_alive = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217677] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.max_header_line = 16384 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.217836] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.secure_proxy_ssl_header = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218058] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.ssl_ca_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218160] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.ssl_cert_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218317] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.ssl_key_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218480] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.tcp_keepidle = 600 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218656] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218820] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] zvm.ca_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.218978] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] zvm.cloud_connector_url = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.219289] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.219463] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] zvm.reachable_timeout = 300 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.219645] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.enforce_new_defaults = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.220041] env[62558]: WARNING oslo_config.cfg [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 466.220234] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.enforce_scope = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.220411] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.policy_default_rule = default {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.220594] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.220768] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.policy_file = policy.yaml {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.220945] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.221121] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.221282] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.221470] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.221697] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.221882] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222076] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222260] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.connection_string = messaging:// {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222430] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.enabled = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222601] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.es_doc_type = notification {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222766] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.es_scroll_size = 10000 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.222936] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.es_scroll_time = 2m {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223116] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.filter_error_trace = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223287] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.hmac_keys = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223455] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.sentinel_service_name = mymaster {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223620] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.socket_timeout = 0.1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223783] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.trace_requests = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.223978] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler.trace_sqlalchemy = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.224180] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler_jaeger.process_tags = {} {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.224351] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler_jaeger.service_name_prefix = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.224509] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] profiler_otlp.service_name_prefix = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.224673] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] remote_debug.host = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.224852] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] remote_debug.port = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225026] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225196] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225358] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225519] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225679] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225838] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.225997] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.226175] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.226391] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.226576] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.226738] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.226906] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227087] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227261] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227432] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227598] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227762] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.227937] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.228116] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.228283] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.228515] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.228706] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.228874] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229057] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229228] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229392] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229555] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229715] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.229885] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230066] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230242] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230414] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230578] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230747] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.230918] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.ssl_version = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.231094] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.231286] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.231507] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_notifications.retry = -1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.231718] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.231902] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_messaging_notifications.transport_url = **** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232090] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.auth_section = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232261] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.auth_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232423] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.cafile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232584] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.certfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232747] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.collect_timing = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.232908] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.connect_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233079] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.connect_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233242] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.endpoint_id = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233400] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.endpoint_override = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233561] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.insecure = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233717] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.keyfile = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.233887] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.max_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234072] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.min_version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234237] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.region_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234401] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.retriable_status_codes = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234560] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.service_name = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234718] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.service_type = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.234879] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.split_loggers = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235048] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.status_code_retries = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235211] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.status_code_retry_delay = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235369] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.timeout = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235527] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.valid_interfaces = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235686] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_limit.version = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.235856] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_reports.file_event_handler = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236032] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236213] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] oslo_reports.log_dir = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236403] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236566] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236726] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.236894] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237073] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237237] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237407] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237568] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237727] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.237893] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238068] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238230] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] vif_plug_ovs_privileged.user = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238437] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.flat_interface = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238626] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238799] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.238970] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239157] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239322] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239486] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239645] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239822] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.239991] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.isolate_vif = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.240174] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.240338] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.240505] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.240671] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.ovsdb_interface = native {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.240830] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] os_vif_ovs.per_port_bridge = False {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241006] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.capabilities = [21] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241178] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241336] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.helper_command = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241497] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241659] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241817] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] privsep_osbrick.user = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.241988] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242166] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.group = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242322] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.helper_command = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242485] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242648] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242804] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] nova_sys_admin.user = None {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 466.242932] env[62558]: DEBUG oslo_service.service [None req-a1043eca-8e9e-4600-8a69-7ec75d09fbaa None None] ******************************************************************************** {{(pid=62558) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 466.243436] env[62558]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 466.746528] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Getting list of instances from cluster (obj){ [ 466.746528] env[62558]: value = "domain-c8" [ 466.746528] env[62558]: _type = "ClusterComputeResource" [ 466.746528] env[62558]: } {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 466.747710] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a980d45e-657b-42cd-afe9-7b4eb791affb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 466.756839] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Got total of 0 instances {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 466.757400] env[62558]: WARNING nova.virt.vmwareapi.driver [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 466.757884] env[62558]: INFO nova.virt.node [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Generated node identity dc830c09-1c36-446a-8af3-d3826bec8b3b [ 466.758153] env[62558]: INFO nova.virt.node [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Wrote node identity dc830c09-1c36-446a-8af3-d3826bec8b3b to /opt/stack/data/n-cpu-1/compute_id [ 467.261091] env[62558]: WARNING nova.compute.manager [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Compute nodes ['dc830c09-1c36-446a-8af3-d3826bec8b3b'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 468.267331] env[62558]: INFO nova.compute.manager [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 469.276018] env[62558]: WARNING nova.compute.manager [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 469.276018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.276018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.276018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.276018] env[62558]: DEBUG nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 469.276018] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a3e71e-f948-42c5-a9c0-f2e666e00832 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.284236] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6c2c9d-fa1b-4a85-bc18-2761a7d9ea7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.298609] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b67475-476e-4fbe-b790-adbcec3aea7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.304813] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6721491b-ef40-4e98-8820-d68a17c11592 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.334880] env[62558]: DEBUG nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181492MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 469.335260] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.335595] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.839395] env[62558]: WARNING nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] No compute node record for cpu-1:dc830c09-1c36-446a-8af3-d3826bec8b3b: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host dc830c09-1c36-446a-8af3-d3826bec8b3b could not be found. [ 470.343025] env[62558]: INFO nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: dc830c09-1c36-446a-8af3-d3826bec8b3b [ 471.853025] env[62558]: DEBUG nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 471.853025] env[62558]: DEBUG nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 472.013219] env[62558]: INFO nova.scheduler.client.report [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] [req-51c00e7b-9f16-4876-9eb5-bf1a8ee5d913] Created resource provider record via placement API for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 472.030060] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af610201-e73d-4311-a220-935b9eaefdf8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.037513] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98258e09-a481-4fc1-8792-81e15a183f4e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.072024] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f2be0d-5716-4e16-85ae-ad6bd7e71e88 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.077877] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ef749b-4707-4d4d-b105-3afe51d8f52f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.093443] env[62558]: DEBUG nova.compute.provider_tree [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 472.631347] env[62558]: DEBUG nova.scheduler.client.report [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 472.631586] env[62558]: DEBUG nova.compute.provider_tree [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 0 to 1 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 472.631729] env[62558]: DEBUG nova.compute.provider_tree [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 472.703149] env[62558]: DEBUG nova.compute.provider_tree [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 1 to 2 during operation: update_traits {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 473.207501] env[62558]: DEBUG nova.compute.resource_tracker [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 473.207879] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.872s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 473.207879] env[62558]: DEBUG nova.service [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Creating RPC server for service compute {{(pid=62558) start /opt/stack/nova/nova/service.py:186}} [ 473.222164] env[62558]: DEBUG nova.service [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] Join ServiceGroup membership for this service compute {{(pid=62558) start /opt/stack/nova/nova/service.py:203}} [ 473.222363] env[62558]: DEBUG nova.servicegroup.drivers.db [None req-e94a9145-5d5d-4f4c-ab7a-fb79728a4d42 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62558) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 509.696503] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.697120] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.203037] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 510.559933] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquiring lock "9b5c5a62-d708-4201-be94-28cff078b2e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.560249] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "9b5c5a62-d708-4201-be94-28cff078b2e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.751526] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.752185] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.753896] env[62558]: INFO nova.compute.claims [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 511.016576] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.017258] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.063853] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.272071] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquiring lock "d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.272480] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.525884] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.597413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.656018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquiring lock "158e242b-3f2d-4cac-b3f3-c51da7f78bb1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.656018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "158e242b-3f2d-4cac-b3f3-c51da7f78bb1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.660773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "d6eaf42d-dd83-4158-a507-9bc71cd1aa43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.660979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "d6eaf42d-dd83-4158-a507-9bc71cd1aa43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.775298] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.913182] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39309cb-45cf-4b81-9a09-0b3b5bf92e5b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.925190] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ff0ccd-59d3-4034-b605-0d7432788fd0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.955674] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e676501e-e366-4acb-9c81-bca813678762 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.964280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ac3024-0f85-4a3e-837c-665cc66f046d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.979996] env[62558]: DEBUG nova.compute.provider_tree [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.054319] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.155408] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 512.164421] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 512.297927] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.486408] env[62558]: DEBUG nova.scheduler.client.report [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.689314] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.690757] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.990808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.992434] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 512.995872] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.399s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.001783] env[62558]: INFO nova.compute.claims [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.511363] env[62558]: DEBUG nova.compute.utils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.513661] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 513.923689] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "99296a9b-151c-4b32-8e6a-ad0e28945aac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.923949] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "99296a9b-151c-4b32-8e6a-ad0e28945aac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.014773] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.208294] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd5aa65-cc02-4106-a1b1-ccbb3987d229 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.217083] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b72150c-4f87-4f94-9da8-e64702332ea8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.252125] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad42c346-9577-4d9f-9a9d-ff5b423f5863 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.262555] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc6e2df-9502-4096-88e1-de7c7ddfeb76 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.267908] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "9f923c30-6c68-4e5d-be8a-4f18e996a999" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.268162] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "9f923c30-6c68-4e5d-be8a-4f18e996a999" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.279388] env[62558]: DEBUG nova.compute.provider_tree [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.428071] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.771561] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.784094] env[62558]: DEBUG nova.scheduler.client.report [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 514.958089] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.033450] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.295574] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.299020] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 515.299020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.245s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.302086] env[62558]: INFO nova.compute.claims [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.308418] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.559119] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.559119] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.559119] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.559329] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.559500] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.559787] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.560245] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.560518] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.561035] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.561417] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.564019] env[62558]: DEBUG nova.virt.hardware [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.564019] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ba6859-0b61-48c6-94aa-91d0c75f444c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.579377] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7049e05-79df-4b9d-bcce-675894209bd3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.613061] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e6ee74-1d8f-4e46-af2e-fbe6bc88d688 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.631522] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 515.641074] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.641387] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0b08710-4f18-4787-b12c-510e11a794e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.655315] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Created folder: OpenStack in parent group-v4. [ 515.655315] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating folder: Project (b841dfe0480c4295b41d903a507f1530). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.655315] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4d549a0-a206-447e-9370-53deef538530 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.665853] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Created folder: Project (b841dfe0480c4295b41d903a507f1530) in parent group-v272451. [ 515.665972] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating folder: Instances. Parent ref: group-v272452. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 515.666197] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5bbbeb1b-b215-4d01-964d-b8e6cec24541 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.678212] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Created folder: Instances in parent group-v272452. [ 515.678674] env[62558]: DEBUG oslo.service.loopingcall [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 515.678888] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 515.679109] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a67b777-4be1-4532-bd62-d6553d9221a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.700314] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 515.700314] env[62558]: value = "task-1266457" [ 515.700314] env[62558]: _type = "Task" [ 515.700314] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.708574] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266457, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.809114] env[62558]: DEBUG nova.compute.utils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.817732] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 515.818859] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 515.916128] env[62558]: DEBUG nova.policy [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28c9b4581e634e46ab8da8b2266ff853', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e9c4a6e7f6d407ea2d2eeaf287dce0f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 515.944915] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "7d66f026-8b3e-4670-ada6-0109d99bb984" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.945222] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "7d66f026-8b3e-4670-ada6-0109d99bb984" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.214254] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266457, 'name': CreateVM_Task, 'duration_secs': 0.34951} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.214512] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 516.215500] env[62558]: DEBUG oslo_vmware.service [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc85a7b-eb48-4013-8191-390c66bd8434 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.222670] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.223023] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.223678] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 516.223976] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3468285b-fc49-4dd5-94db-b0ee7b8f3ba6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.230029] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 516.230029] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e5b924-74de-bb1a-49ba-f41146e712a9" [ 516.230029] env[62558]: _type = "Task" [ 516.230029] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.238237] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e5b924-74de-bb1a-49ba-f41146e712a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.322882] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 516.334557] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Successfully created port: b080e1c5-09b1-467b-a528-5319a1ab4c75 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.448422] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.503072] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9fa331-1eb1-47af-8fea-e513a048e301 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.511973] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50da4bb8-0554-4538-97c8-fa4a7f24c186 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.547831] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8a3197-42a8-4e10-bd12-2a125c86387e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.555339] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a210db7d-a7b1-4f66-8d1a-9c0609ad5832 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.569223] env[62558]: DEBUG nova.compute.provider_tree [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.746354] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.747358] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 516.747759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.747856] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.748354] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 516.748799] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de98d25b-8fea-4e4b-a6ef-0e6a60ef8488 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.770889] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 516.771160] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 516.772447] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb55f6a-9567-46ef-9a73-d843d8ccbe62 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.785191] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6744080d-da37-465f-a884-2dae803e55c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.789560] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 516.789560] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cb31a2-c2bb-6583-99ec-8811af38223a" [ 516.789560] env[62558]: _type = "Task" [ 516.789560] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.800987] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cb31a2-c2bb-6583-99ec-8811af38223a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.973490] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.076110] env[62558]: DEBUG nova.scheduler.client.report [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.223759] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.301058] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Preparing fetch location {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 517.301668] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating directory with path [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 517.301839] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb99c82b-0d63-44ac-a219-4eb2cf1ac411 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.321713] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Created directory with path [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 517.321713] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Fetch image to [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 517.321713] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Downloading image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk on the data store datastore2 {{(pid=62558) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 517.322629] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f24c47-e1db-49c0-9561-2a02d258251d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.329811] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4efa1b81-f340-42c7-9fe5-3086dec453ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.338990] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 517.342262] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a9a61f-553a-4c58-a2b9-b7ab69d44e0a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.388113] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e6161a-8fc5-40f5-a7c9-b686a88b67a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.395123] env[62558]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ebc2d178-ca93-4a21-85b2-b0b5da513740 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.399419] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.400373] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.400543] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.400724] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.400888] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.401057] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.401268] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.401425] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.401584] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.401741] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.401909] env[62558]: DEBUG nova.virt.hardware [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.402708] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3731f958-f8e5-4032-a1b9-adc13b4d2b9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.419089] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b323b8-a492-4cec-ab86-5ccda6198a0b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.423435] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Downloading image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to the data store datastore2 {{(pid=62558) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 517.582525] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.583156] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.587506] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.290s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.589519] env[62558]: INFO nova.compute.claims [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 517.607128] env[62558]: DEBUG oslo_vmware.rw_handles [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62558) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 517.729991] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Getting list of instances from cluster (obj){ [ 517.729991] env[62558]: value = "domain-c8" [ 517.729991] env[62558]: _type = "ClusterComputeResource" [ 517.729991] env[62558]: } {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 517.731672] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4e1f4f-3188-4376-9e80-aed4c36f7d4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.747593] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Got total of 1 instances {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 517.748369] env[62558]: WARNING nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] While synchronizing instance power states, found 3 instances in the database and 1 instances on the hypervisor. [ 517.748662] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 06df57a7-2c53-4f79-bec4-e46cfe1dca63 {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 517.748964] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 9b5c5a62-d708-4201-be94-28cff078b2e8 {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 517.749183] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid d02e6383-42d9-40db-89a3-ba8b5595b95b {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 517.751347] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.751917] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "9b5c5a62-d708-4201-be94-28cff078b2e8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.752020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.752179] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.752722] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Getting list of instances from cluster (obj){ [ 517.752722] env[62558]: value = "domain-c8" [ 517.752722] env[62558]: _type = "ClusterComputeResource" [ 517.752722] env[62558]: } {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 517.753745] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e878f94-8a4f-4f43-b466-651a5e3d015d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.765961] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Got total of 1 instances {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 518.030875] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquiring lock "b00d6153-c94f-4605-9b97-9a3a813ee93b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.031167] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "b00d6153-c94f-4605-9b97-9a3a813ee93b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.094670] env[62558]: DEBUG nova.compute.utils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.098374] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 518.134631] env[62558]: ERROR nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 518.134631] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.134631] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.134631] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.134631] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.134631] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.134631] env[62558]: ERROR nova.compute.manager raise self.value [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.134631] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 518.134631] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.134631] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 518.135336] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.135336] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 518.135336] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 518.135336] env[62558]: ERROR nova.compute.manager [ 518.135336] env[62558]: Traceback (most recent call last): [ 518.135336] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 518.135336] env[62558]: listener.cb(fileno) [ 518.135336] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.135336] env[62558]: result = function(*args, **kwargs) [ 518.135336] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.135336] env[62558]: return func(*args, **kwargs) [ 518.135336] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.135336] env[62558]: raise e [ 518.135336] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.135336] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 518.135336] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.135336] env[62558]: created_port_ids = self._update_ports_for_instance( [ 518.135336] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.135336] env[62558]: with excutils.save_and_reraise_exception(): [ 518.135336] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.135336] env[62558]: self.force_reraise() [ 518.135336] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.135336] env[62558]: raise self.value [ 518.135336] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.135336] env[62558]: updated_port = self._update_port( [ 518.135336] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.135336] env[62558]: _ensure_no_port_binding_failure(port) [ 518.135336] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.135336] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 518.136455] env[62558]: nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 518.136455] env[62558]: Removing descriptor: 15 [ 518.136455] env[62558]: ERROR nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Traceback (most recent call last): [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] yield resources [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.driver.spawn(context, instance, image_meta, [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.136455] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] vm_ref = self.build_virtual_machine(instance, [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] for vif in network_info: [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self._sync_wrapper(fn, *args, **kwargs) [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.wait() [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self[:] = self._gt.wait() [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self._exit_event.wait() [ 518.136799] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] result = hub.switch() [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self.greenlet.switch() [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] result = function(*args, **kwargs) [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return func(*args, **kwargs) [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise e [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] nwinfo = self.network_api.allocate_for_instance( [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.137234] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] created_port_ids = self._update_ports_for_instance( [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] with excutils.save_and_reraise_exception(): [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.force_reraise() [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise self.value [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] updated_port = self._update_port( [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] _ensure_no_port_binding_failure(port) [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.137590] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise exception.PortBindingFailed(port_id=port['id']) [ 518.138588] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 518.138588] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] [ 518.138588] env[62558]: INFO nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Terminating instance [ 518.145380] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquiring lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.145476] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquired lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.148476] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 518.377873] env[62558]: DEBUG oslo_vmware.rw_handles [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Completed reading data from the image iterator. {{(pid=62558) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 518.378833] env[62558]: DEBUG oslo_vmware.rw_handles [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 518.435059] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Downloaded image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk on the data store datastore2 {{(pid=62558) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 518.435059] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Caching image {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 518.435059] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Copying Virtual Disk [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk to [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 518.436329] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51f3c203-d43d-4b76-9114-1cdeff38658c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.446425] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 518.446425] env[62558]: value = "task-1266458" [ 518.446425] env[62558]: _type = "Task" [ 518.446425] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.456030] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.537035] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.599645] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.688821] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 518.800148] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319c2106-f1a4-4630-ba77-fc45ea465cb9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.811867] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf80d31-0387-45d0-ba5f-3a34fc682bbb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.850261] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c80a6a1-bce9-46df-bfe0-eaacf28a529c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.866197] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20479009-7f70-4250-82ad-5c63c6590342 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.878999] env[62558]: DEBUG nova.compute.provider_tree [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.961203] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.969183] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266458, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.070157] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.383324] env[62558]: DEBUG nova.scheduler.client.report [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.458163] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718041} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.458163] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Copied Virtual Disk [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk to [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 519.458163] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleting the datastore file [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 519.459049] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ce25638-cb27-4fb0-a997-bedd074f1226 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.464567] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Releasing lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.464989] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 519.465205] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 519.467487] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8182abf-4a02-48dc-86d5-e2f87e92618f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.471117] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 519.471117] env[62558]: value = "task-1266459" [ 519.471117] env[62558]: _type = "Task" [ 519.471117] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.482642] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad1f5e3-0fb5-48af-ac7d-f3255dacc417 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.505142] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.523387] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b5c5a62-d708-4201-be94-28cff078b2e8 could not be found. [ 519.523629] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 519.524113] env[62558]: INFO nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Took 0.06 seconds to destroy the instance on the hypervisor. [ 519.524375] env[62558]: DEBUG oslo.service.loopingcall [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 519.524596] env[62558]: DEBUG nova.compute.manager [-] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 519.524687] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 519.558516] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 519.610502] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.641723] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.642012] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.642178] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.642360] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.642504] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.642694] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.642856] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.643015] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.643186] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.643341] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.643502] env[62558]: DEBUG nova.virt.hardware [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.644364] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d3383a-fc3a-4988-8b8a-2a1886cd953d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.652740] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d2d01b-1698-4a35-ab8d-1f09e4dd9b5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.665982] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 519.671994] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating folder: Project (f2d0b0da015442179c0a9915ae0caefd). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 519.672177] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d96d876-b59e-4bfd-a662-e32357497cb2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.683687] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Created folder: Project (f2d0b0da015442179c0a9915ae0caefd) in parent group-v272451. [ 519.683985] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating folder: Instances. Parent ref: group-v272455. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 519.684135] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-020b644d-c715-4b19-874a-6827e56dbc9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.693773] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Created folder: Instances in parent group-v272455. [ 519.694028] env[62558]: DEBUG oslo.service.loopingcall [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 519.694224] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 519.694429] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87292072-a674-4555-9518-d8d03b07db07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.711843] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 519.711843] env[62558]: value = "task-1266462" [ 519.711843] env[62558]: _type = "Task" [ 519.711843] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.719866] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266462, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.889717] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.890380] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.896251] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.207s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.898259] env[62558]: INFO nova.compute.claims [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.984186] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024255} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.984502] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 519.984711] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Moving file from [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5/bd09177c-5fd7-495c-893b-5ec219b6b3de to [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de. {{(pid=62558) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 519.984998] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a59873c5-9353-4b3a-910e-295312fe6a5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.996368] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 519.996368] env[62558]: value = "task-1266463" [ 519.996368] env[62558]: _type = "Task" [ 519.996368] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.007016] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266463, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.035382] env[62558]: DEBUG nova.compute.manager [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Received event network-changed-b080e1c5-09b1-467b-a528-5319a1ab4c75 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 520.036234] env[62558]: DEBUG nova.compute.manager [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Refreshing instance network info cache due to event network-changed-b080e1c5-09b1-467b-a528-5319a1ab4c75. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 520.036234] env[62558]: DEBUG oslo_concurrency.lockutils [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] Acquiring lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.036234] env[62558]: DEBUG oslo_concurrency.lockutils [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] Acquired lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.036417] env[62558]: DEBUG nova.network.neutron [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Refreshing network info cache for port b080e1c5-09b1-467b-a528-5319a1ab4c75 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 520.061180] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.227275] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266462, 'name': CreateVM_Task, 'duration_secs': 0.365421} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.227460] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 520.228478] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.228740] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.229964] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 520.229964] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63e921da-45ce-4a15-99aa-790160115663 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.235741] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 520.235741] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521fb063-5915-fa45-acf3-7970b1f4c55c" [ 520.235741] env[62558]: _type = "Task" [ 520.235741] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.248505] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521fb063-5915-fa45-acf3-7970b1f4c55c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.408931] env[62558]: DEBUG nova.compute.utils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.414095] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 520.414095] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 520.513022] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266463, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.042477} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.513022] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] File moved {{(pid=62558) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 520.513022] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Cleaning up location [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5 {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 520.513022] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleting the datastore file [datastore2] vmware_temp/7aff1e02-e0a7-48eb-b7e0-284191f1edf5 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 520.513022] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48d49acb-81c1-46e7-b0e0-494d3a91787b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.517676] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 520.517676] env[62558]: value = "task-1266464" [ 520.517676] env[62558]: _type = "Task" [ 520.517676] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.526812] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.530146] env[62558]: DEBUG nova.policy [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91a47c92b8df40049fd2a8f3ca70b1c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '279b995323cf405493f63e8444306e30', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.564142] env[62558]: INFO nova.compute.manager [-] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Took 1.04 seconds to deallocate network for instance. [ 520.570555] env[62558]: DEBUG nova.compute.claims [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 520.570775] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.704683] env[62558]: DEBUG nova.network.neutron [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.748409] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521fb063-5915-fa45-acf3-7970b1f4c55c, 'name': SearchDatastore_Task, 'duration_secs': 0.008452} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 520.748926] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.749205] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 520.749420] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.897093] env[62558]: DEBUG nova.network.neutron [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.912489] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.033885] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025752} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.034143] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 521.034854] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c33c8a8-9e57-4336-bae4-561938202520 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.043019] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 521.043019] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e29745-6861-8025-5d02-e988a9aa0589" [ 521.043019] env[62558]: _type = "Task" [ 521.043019] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.048406] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Successfully created port: 90d1f7f3-3524-4853-a84d-8f3ed8fbce17 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.056879] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e29745-6861-8025-5d02-e988a9aa0589, 'name': SearchDatastore_Task, 'duration_secs': 0.00984} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.060299] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.060299] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 06df57a7-2c53-4f79-bec4-e46cfe1dca63/06df57a7-2c53-4f79-bec4-e46cfe1dca63.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 521.060299] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.060299] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 521.060479] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51b76abc-3b95-4e28-a946-3358ecbce679 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.060874] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f97d7400-162a-480b-b129-a53be6d0f51f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.069461] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 521.069461] env[62558]: value = "task-1266465" [ 521.069461] env[62558]: _type = "Task" [ 521.069461] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.073989] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 521.074209] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 521.075299] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d0b97d2-e94b-49d6-9195-a23cb95d6b7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.080416] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.085331] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 521.085331] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5241e985-b222-68c2-b991-544d8523a58a" [ 521.085331] env[62558]: _type = "Task" [ 521.085331] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.092479] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5241e985-b222-68c2-b991-544d8523a58a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.128144] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceca27f7-1a0a-4096-8fba-dec95d155e33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.137318] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1aaa98-ef9c-4bef-af84-c2948fa2a1b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.173614] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac8a19d-413b-4860-86e3-ec5609dc2e1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.182585] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f98ade3-124a-44d8-a012-857fa78c1884 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.197196] env[62558]: DEBUG nova.compute.provider_tree [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.400430] env[62558]: DEBUG oslo_concurrency.lockutils [req-37b25122-fb7f-443c-a716-5ff547d34a25 req-2c5c2090-30d5-4e91-b2f7-f6276c139d79 service nova] Releasing lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.586419] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474827} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.586419] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 06df57a7-2c53-4f79-bec4-e46cfe1dca63/06df57a7-2c53-4f79-bec4-e46cfe1dca63.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 521.587089] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 521.591196] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d504f2eb-c749-4710-aa08-089bfed241a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.599791] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5241e985-b222-68c2-b991-544d8523a58a, 'name': SearchDatastore_Task, 'duration_secs': 0.007581} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.602167] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 521.602167] env[62558]: value = "task-1266466" [ 521.602167] env[62558]: _type = "Task" [ 521.602167] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.603083] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c87d390b-1099-4c04-ae2f-d0a895953556 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.617491] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266466, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.617491] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 521.617491] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f7b96-149b-451c-a6c0-452249ceda99" [ 521.617491] env[62558]: _type = "Task" [ 521.617491] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.627564] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f7b96-149b-451c-a6c0-452249ceda99, 'name': SearchDatastore_Task, 'duration_secs': 0.008517} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.627867] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.628187] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 521.628510] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49591bc0-619e-4791-a179-fce8776996eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.636117] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 521.636117] env[62558]: value = "task-1266467" [ 521.636117] env[62558]: _type = "Task" [ 521.636117] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.646208] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.703756] env[62558]: DEBUG nova.scheduler.client.report [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.925240] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 521.960437] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 521.960682] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 521.962618] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 521.962618] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 521.962618] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 521.962618] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 521.962618] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 521.962965] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 521.962965] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 521.962965] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 521.962965] env[62558]: DEBUG nova.virt.hardware [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 521.964396] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4626cc-0486-4213-a255-bb2d55d225cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.977334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cb81bc-7be3-45c1-9f92-9b74686a3b4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.999207] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquiring lock "3c173701-9f61-4561-ac3c-7ac3de4101d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.000115] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "3c173701-9f61-4561-ac3c-7ac3de4101d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.121110] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266466, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063011} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.121110] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 522.121249] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0b3594-0837-4743-b869-84824f02b9fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.148578] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 06df57a7-2c53-4f79-bec4-e46cfe1dca63/06df57a7-2c53-4f79-bec4-e46cfe1dca63.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 522.153402] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.154095] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31692faf-ce59-4ac2-880f-8279fe9ddd86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.169436] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.169436] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 522.169436] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 522.175584] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473463} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.176928] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 522.177199] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 522.177555] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 522.177555] env[62558]: value = "task-1266468" [ 522.177555] env[62558]: _type = "Task" [ 522.177555] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.177840] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf245264-c087-4fe7-828f-72bb8a75e896 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.188326] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266468, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.190329] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 522.190329] env[62558]: value = "task-1266469" [ 522.190329] env[62558]: _type = "Task" [ 522.190329] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.198648] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.210933] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.211448] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.214366] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.524s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.219540] env[62558]: INFO nova.compute.claims [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.629401] env[62558]: ERROR nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 522.629401] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.629401] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.629401] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.629401] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.629401] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.629401] env[62558]: ERROR nova.compute.manager raise self.value [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.629401] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 522.629401] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.629401] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 522.630451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.630451] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 522.630451] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 522.630451] env[62558]: ERROR nova.compute.manager [ 522.630451] env[62558]: Traceback (most recent call last): [ 522.630451] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 522.630451] env[62558]: listener.cb(fileno) [ 522.630451] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.630451] env[62558]: result = function(*args, **kwargs) [ 522.630451] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.630451] env[62558]: return func(*args, **kwargs) [ 522.630451] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.630451] env[62558]: raise e [ 522.630451] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.630451] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 522.630451] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.630451] env[62558]: created_port_ids = self._update_ports_for_instance( [ 522.630451] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.630451] env[62558]: with excutils.save_and_reraise_exception(): [ 522.630451] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.630451] env[62558]: self.force_reraise() [ 522.630451] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.630451] env[62558]: raise self.value [ 522.630451] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.630451] env[62558]: updated_port = self._update_port( [ 522.630451] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.630451] env[62558]: _ensure_no_port_binding_failure(port) [ 522.630451] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.630451] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 522.631339] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 522.631339] env[62558]: Removing descriptor: 15 [ 522.631339] env[62558]: ERROR nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Traceback (most recent call last): [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] yield resources [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.driver.spawn(context, instance, image_meta, [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.631339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] vm_ref = self.build_virtual_machine(instance, [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] for vif in network_info: [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self._sync_wrapper(fn, *args, **kwargs) [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.wait() [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self[:] = self._gt.wait() [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self._exit_event.wait() [ 522.631726] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] result = hub.switch() [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self.greenlet.switch() [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] result = function(*args, **kwargs) [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return func(*args, **kwargs) [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise e [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] nwinfo = self.network_api.allocate_for_instance( [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.632505] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] created_port_ids = self._update_ports_for_instance( [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] with excutils.save_and_reraise_exception(): [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.force_reraise() [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise self.value [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] updated_port = self._update_port( [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] _ensure_no_port_binding_failure(port) [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.633022] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise exception.PortBindingFailed(port_id=port['id']) [ 522.633339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 522.633339] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] [ 522.633339] env[62558]: INFO nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Terminating instance [ 522.633339] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquiring lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.633339] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquired lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.633339] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.672573] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 522.673120] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 522.673312] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 522.673442] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 522.673609] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 522.673687] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Didn't find any instances for network info cache update. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 522.674206] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.674453] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.674646] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.674828] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.675099] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.675327] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.675494] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 522.675641] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.691218] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266468, 'name': ReconfigVM_Task, 'duration_secs': 0.404082} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.695418] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 06df57a7-2c53-4f79-bec4-e46cfe1dca63/06df57a7-2c53-4f79-bec4-e46cfe1dca63.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 522.696399] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8350f4e3-f0b7-4443-b8aa-5224e44543da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.705448] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266469, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.706950] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 522.706950] env[62558]: value = "task-1266470" [ 522.706950] env[62558]: _type = "Task" [ 522.706950] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.717859] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266470, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.728023] env[62558]: DEBUG nova.compute.utils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.731812] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.731812] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 522.806162] env[62558]: DEBUG nova.policy [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36729b6b4df2408f98374f00a6b34e96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ab6d633c1cd4abaa63566ab5a0b2429', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.155650] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Successfully created port: 346a4760-dcb6-456b-8249-a1c7fa4c00a7 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.159037] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.178735] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.203738] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266469, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.544998} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.204065] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 523.204924] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5548c3be-8233-46e0-9901-e82f6ceadf6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.237455] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 523.237952] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266470, 'name': Rename_Task, 'duration_secs': 0.130963} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.238530] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.245114] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef2f256f-62a3-430b-8856-0ec9e4ad9bef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.261694] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 523.262540] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5766f14-a57f-4881-9080-b9efdd17db97 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.270981] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 523.270981] env[62558]: value = "task-1266472" [ 523.270981] env[62558]: _type = "Task" [ 523.270981] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.272255] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 523.272255] env[62558]: value = "task-1266471" [ 523.272255] env[62558]: _type = "Task" [ 523.272255] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.273405] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.292299] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266471, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.501631] env[62558]: DEBUG nova.compute.manager [req-36272d67-01d2-484a-a219-14ce2ed66e2e req-c9e0cd87-bd8a-4d95-893f-15d497f8610e service nova] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Received event network-vif-deleted-b080e1c5-09b1-467b-a528-5319a1ab4c75 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 523.560215] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb38345-222e-4d3a-98ee-3ac42c6eeab3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.569450] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df1be90-5135-469b-8979-8a4b82b3ea94 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.599525] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32734f23-f0f9-4f10-aa84-59c1396829b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.607236] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b14a50a-857b-4139-a5b0-ccd810f44463 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.620760] env[62558]: DEBUG nova.compute.provider_tree [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.780310] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Releasing lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.780310] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 523.780310] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 523.788786] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0aab88d8-2362-494f-96c2-4bab1bdf1392 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.790524] env[62558]: DEBUG oslo_vmware.api [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266472, 'name': PowerOnVM_Task, 'duration_secs': 0.45272} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.790860] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 523.791059] env[62558]: INFO nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Took 8.76 seconds to spawn the instance on the hypervisor. [ 523.791309] env[62558]: DEBUG nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 523.793707] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a7c73e-22dc-4a28-926d-9762cae4d3b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.801081] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266471, 'name': ReconfigVM_Task, 'duration_secs': 0.271971} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.803104] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 523.806645] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753d1642-16cb-4f88-8831-3b03209043b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.819014] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6231124a-4811-4487-868a-f78c9a0f3726 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.836890] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 523.836890] env[62558]: value = "task-1266473" [ 523.836890] env[62558]: _type = "Task" [ 523.836890] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.837482] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d could not be found. [ 523.837694] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 523.837921] env[62558]: INFO nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 523.838203] env[62558]: DEBUG oslo.service.loopingcall [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.838937] env[62558]: DEBUG nova.compute.manager [-] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 523.839059] env[62558]: DEBUG nova.network.neutron [-] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.849887] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266473, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.885023] env[62558]: DEBUG nova.network.neutron [-] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.126072] env[62558]: DEBUG nova.scheduler.client.report [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.272535] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.309183] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.310023] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.310023] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.310023] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.310023] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.310236] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.312031] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.312031] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.312031] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.312031] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.312031] env[62558]: DEBUG nova.virt.hardware [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.315457] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61a5838-da1c-4679-b9a0-936c272b2cf9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.328881] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4364b83-bd32-4939-9da4-3335c81cbd65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.369892] env[62558]: INFO nova.compute.manager [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Took 13.66 seconds to build instance. [ 524.370884] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266473, 'name': Rename_Task, 'duration_secs': 0.38888} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.371445] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 524.371884] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0afc6c9f-e3ab-4d35-a4c8-d67a2aa1bb0c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.382652] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 524.382652] env[62558]: value = "task-1266474" [ 524.382652] env[62558]: _type = "Task" [ 524.382652] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.387863] env[62558]: DEBUG nova.network.neutron [-] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.398567] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266474, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.635154] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.635519] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.638928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.681s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.641226] env[62558]: INFO nova.compute.claims [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.872809] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6a4064e9-39a3-4a06-b68f-eaac0e81998b tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.176s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.874227] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.123s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.874414] env[62558]: INFO nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] During sync_power_state the instance has a pending task (spawning). Skip. [ 524.874703] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.891916] env[62558]: INFO nova.compute.manager [-] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Took 1.05 seconds to deallocate network for instance. [ 524.898449] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266474, 'name': PowerOnVM_Task} progress is 98%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.898997] env[62558]: DEBUG nova.compute.claims [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 524.899214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.926242] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "b08cb5a2-56d1-43da-ae4a-684077dd517f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.926242] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "b08cb5a2-56d1-43da-ae4a-684077dd517f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.146625] env[62558]: DEBUG nova.compute.utils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.150598] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.150793] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.319466] env[62558]: DEBUG nova.policy [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02465b858daa49858a0125be964cd35e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0f0bd91a1ca41a891eddcf891ecbbdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.346023] env[62558]: ERROR nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 525.346023] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.346023] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.346023] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.346023] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.346023] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.346023] env[62558]: ERROR nova.compute.manager raise self.value [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.346023] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 525.346023] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.346023] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 525.346533] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.346533] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 525.346533] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 525.346533] env[62558]: ERROR nova.compute.manager [ 525.346767] env[62558]: Traceback (most recent call last): [ 525.346802] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 525.346802] env[62558]: listener.cb(fileno) [ 525.346802] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.346802] env[62558]: result = function(*args, **kwargs) [ 525.346802] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.346802] env[62558]: return func(*args, **kwargs) [ 525.346802] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.346802] env[62558]: raise e [ 525.347017] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.347017] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 525.347017] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.347017] env[62558]: created_port_ids = self._update_ports_for_instance( [ 525.347017] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.347017] env[62558]: with excutils.save_and_reraise_exception(): [ 525.347017] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.347017] env[62558]: self.force_reraise() [ 525.347017] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.347017] env[62558]: raise self.value [ 525.347017] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.347017] env[62558]: updated_port = self._update_port( [ 525.347017] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.347017] env[62558]: _ensure_no_port_binding_failure(port) [ 525.347017] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.347017] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 525.347017] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 525.347017] env[62558]: Removing descriptor: 15 [ 525.347856] env[62558]: ERROR nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Traceback (most recent call last): [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] yield resources [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.driver.spawn(context, instance, image_meta, [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] vm_ref = self.build_virtual_machine(instance, [ 525.347856] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] for vif in network_info: [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self._sync_wrapper(fn, *args, **kwargs) [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.wait() [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self[:] = self._gt.wait() [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self._exit_event.wait() [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.348283] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] result = hub.switch() [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self.greenlet.switch() [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] result = function(*args, **kwargs) [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return func(*args, **kwargs) [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise e [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] nwinfo = self.network_api.allocate_for_instance( [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] created_port_ids = self._update_ports_for_instance( [ 525.348623] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] with excutils.save_and_reraise_exception(): [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.force_reraise() [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise self.value [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] updated_port = self._update_port( [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] _ensure_no_port_binding_failure(port) [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise exception.PortBindingFailed(port_id=port['id']) [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 525.348984] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] [ 525.349384] env[62558]: INFO nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Terminating instance [ 525.351013] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.351197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquired lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.351368] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 525.377357] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.396767] env[62558]: DEBUG oslo_vmware.api [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266474, 'name': PowerOnVM_Task, 'duration_secs': 0.687507} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.396894] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 525.397543] env[62558]: INFO nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Took 5.79 seconds to spawn the instance on the hypervisor. [ 525.397771] env[62558]: DEBUG nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 525.398566] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b930ae0-1d9c-47ed-8504-afc7a2f7d0e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.652297] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.879563] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e26dee-918f-46fc-aeb2-f55b66274904 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.890952] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.899022] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe1ade9-750d-47b4-89c0-f01a06896a0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.903302] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.945690] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec13739d-7755-49d7-adb1-88506016e601 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.949129] env[62558]: INFO nova.compute.manager [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Took 13.92 seconds to build instance. [ 525.956867] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a5e8e0-f266-453f-a910-897e77367d57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.975791] env[62558]: DEBUG nova.compute.provider_tree [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.029587] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.262195] env[62558]: DEBUG nova.compute.manager [None req-991b8f18-6fbb-415b-a094-842dd1d675fa tempest-ServerDiagnosticsV248Test-857757544 tempest-ServerDiagnosticsV248Test-857757544-project-admin] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 526.263942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac96996-a4c5-4906-a28a-3b5aa7acaf35 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.273330] env[62558]: INFO nova.compute.manager [None req-991b8f18-6fbb-415b-a094-842dd1d675fa tempest-ServerDiagnosticsV248Test-857757544 tempest-ServerDiagnosticsV248Test-857757544-project-admin] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Retrieving diagnostics [ 526.274799] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28675ae4-3e15-42f6-8ad5-c3c32036ce98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.451642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-60e6088b-b641-4d32-88a4-cd8421111713 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.434s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.452972] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.701s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.455614] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba14db74-8e97-4f0f-96f1-6968d619f7db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.482084] env[62558]: DEBUG nova.scheduler.client.report [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.531838] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Releasing lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.532317] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 526.532525] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 526.532791] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bbb09aa-fb92-438a-b4b3-0a5e310dcf90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.542661] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7770d1f1-8e1e-4dfb-ad01-abc803724b1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.575377] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6eaf42d-dd83-4158-a507-9bc71cd1aa43 could not be found. [ 526.575377] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 526.575377] env[62558]: INFO nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Took 0.04 seconds to destroy the instance on the hypervisor. [ 526.575377] env[62558]: DEBUG oslo.service.loopingcall [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.575377] env[62558]: DEBUG nova.compute.manager [-] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.575377] env[62558]: DEBUG nova.network.neutron [-] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 526.598040] env[62558]: DEBUG nova.network.neutron [-] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.607280] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Received event network-changed-90d1f7f3-3524-4853-a84d-8f3ed8fbce17 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.607280] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Refreshing instance network info cache due to event network-changed-90d1f7f3-3524-4853-a84d-8f3ed8fbce17. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.607280] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Acquiring lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.607280] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Acquired lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.607280] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Refreshing network info cache for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 526.665627] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 526.773260] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.773260] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.773260] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.773443] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.773443] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.773524] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.773748] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.773932] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.774216] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.774473] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.774751] env[62558]: DEBUG nova.virt.hardware [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.775848] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbc7ca6-9c6a-4dfd-9a47-fac9a9f73b38 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.788241] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5599cdf-9a1e-42dc-9958-f08a0369e963 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.962588] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.975598] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.523s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.033020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.033020] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.033020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.705s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.033020] env[62558]: INFO nova.compute.claims [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.076453] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Successfully created port: 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.100252] env[62558]: DEBUG nova.network.neutron [-] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.168065] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.307632] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.503828] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.537142] env[62558]: DEBUG nova.compute.utils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.537142] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.537275] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 527.602944] env[62558]: INFO nova.compute.manager [-] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Took 1.03 seconds to deallocate network for instance. [ 527.605698] env[62558]: DEBUG nova.compute.claims [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.605878] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.612202] env[62558]: DEBUG nova.policy [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8f49d29496540a883611670b2fafee4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa1d90fab22044528de22ba940b9a2a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.810749] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Releasing lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.811054] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Received event network-vif-deleted-90d1f7f3-3524-4853-a84d-8f3ed8fbce17 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.811333] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Received event network-changed-346a4760-dcb6-456b-8249-a1c7fa4c00a7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.811431] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Refreshing instance network info cache due to event network-changed-346a4760-dcb6-456b-8249-a1c7fa4c00a7. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.812830] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Acquiring lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.812830] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Acquired lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.812830] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Refreshing network info cache for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 528.042101] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.253052] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea1f9b3-3eff-4a4b-918f-60d84fae88cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.260995] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed3b1cf-4e48-4524-bdcc-5c07aa6c275f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.293355] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce89e58a-15fe-4316-af3d-12eac507b173 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.301827] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2207f2a-67eb-4646-8161-1fbba4226a21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.319558] env[62558]: DEBUG nova.compute.provider_tree [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.345648] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.576895] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Successfully created port: a8a4a427-c551-4893-abfe-62502b097be8 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.676727] env[62558]: DEBUG nova.network.neutron [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.823027] env[62558]: DEBUG nova.scheduler.client.report [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.053989] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.079557] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.079812] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.079967] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.080161] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.080307] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.080475] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.081580] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.081834] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.081927] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.082203] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.082287] env[62558]: DEBUG nova.virt.hardware [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.083305] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e8d52d-84d0-4fa4-8cc6-912ed6623507 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.093863] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfe6817-41f6-45f3-a215-c3c8780ccbc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.179884] env[62558]: DEBUG oslo_concurrency.lockutils [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] Releasing lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.180208] env[62558]: DEBUG nova.compute.manager [req-96589e72-304e-4700-a82f-e2696556fc34 req-ccc698ca-31d4-4069-8e12-79fd15b5ad09 service nova] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Received event network-vif-deleted-346a4760-dcb6-456b-8249-a1c7fa4c00a7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.330729] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.331357] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.335195] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.362s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.336738] env[62558]: INFO nova.compute.claims [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.845842] env[62558]: DEBUG nova.compute.utils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.852237] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.852393] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.854522] env[62558]: INFO nova.compute.manager [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Rebuilding instance [ 529.921035] env[62558]: DEBUG nova.compute.manager [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 529.924025] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be8b402-8b6e-4214-8cb0-99ce417e8de6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.101587] env[62558]: DEBUG nova.policy [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8434f838ad104f9491f9cf4df7093672', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49dd92c04a6b48c5a71b889d5e016135', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.354988] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.435758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "9316facd-0c95-4720-b0d3-c0a39df81619" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.435978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.440564] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 530.441297] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8a97dd0-30cd-4bf5-ad37-0a50a50d8cfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.453754] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 530.453754] env[62558]: value = "task-1266475" [ 530.453754] env[62558]: _type = "Task" [ 530.453754] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.472422] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.615315] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5532c599-7d9d-4f8d-abda-924c362d74cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.624801] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a459ab7e-cad3-47cc-b83e-f8d23a54500f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.660114] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5bccaf-e6ba-4192-81e0-b191d39f7ca4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.667330] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255e3b9c-7f1d-499d-ad47-254d5114a5b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.684267] env[62558]: DEBUG nova.compute.provider_tree [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.971779] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266475, 'name': PowerOffVM_Task, 'duration_secs': 0.125751} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.972106] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 530.972368] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 530.973168] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d6f9a8-3946-4208-b68b-1578156c37cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.981113] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 530.981495] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea382c6d-0c5c-4036-bee0-e58f597e86d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.004663] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 531.004875] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 531.005054] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleting the datastore file [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 531.005382] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2266c8e8-ec91-465d-aec2-10a4ace64ee9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.012304] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 531.012304] env[62558]: value = "task-1266477" [ 531.012304] env[62558]: _type = "Task" [ 531.012304] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.024074] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266477, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.192035] env[62558]: DEBUG nova.scheduler.client.report [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.224372] env[62558]: ERROR nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 531.224372] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.224372] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.224372] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.224372] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.224372] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.224372] env[62558]: ERROR nova.compute.manager raise self.value [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.224372] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 531.224372] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.224372] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 531.224858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.224858] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 531.224858] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 531.224858] env[62558]: ERROR nova.compute.manager [ 531.224858] env[62558]: Traceback (most recent call last): [ 531.224858] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 531.224858] env[62558]: listener.cb(fileno) [ 531.224858] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.224858] env[62558]: result = function(*args, **kwargs) [ 531.224858] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.224858] env[62558]: return func(*args, **kwargs) [ 531.224858] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 531.224858] env[62558]: raise e [ 531.224858] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.224858] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 531.224858] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.224858] env[62558]: created_port_ids = self._update_ports_for_instance( [ 531.224858] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.224858] env[62558]: with excutils.save_and_reraise_exception(): [ 531.224858] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.224858] env[62558]: self.force_reraise() [ 531.224858] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.224858] env[62558]: raise self.value [ 531.224858] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.224858] env[62558]: updated_port = self._update_port( [ 531.224858] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.224858] env[62558]: _ensure_no_port_binding_failure(port) [ 531.224858] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.224858] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 531.226529] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 531.226529] env[62558]: Removing descriptor: 18 [ 531.226529] env[62558]: ERROR nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Traceback (most recent call last): [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] yield resources [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.driver.spawn(context, instance, image_meta, [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 531.226529] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] vm_ref = self.build_virtual_machine(instance, [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] vif_infos = vmwarevif.get_vif_info(self._session, [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] for vif in network_info: [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self._sync_wrapper(fn, *args, **kwargs) [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.wait() [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self[:] = self._gt.wait() [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self._exit_event.wait() [ 531.227561] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] result = hub.switch() [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self.greenlet.switch() [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] result = function(*args, **kwargs) [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return func(*args, **kwargs) [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise e [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] nwinfo = self.network_api.allocate_for_instance( [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 531.227907] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] created_port_ids = self._update_ports_for_instance( [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] with excutils.save_and_reraise_exception(): [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.force_reraise() [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise self.value [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] updated_port = self._update_port( [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] _ensure_no_port_binding_failure(port) [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.228265] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise exception.PortBindingFailed(port_id=port['id']) [ 531.228571] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 531.228571] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] [ 531.228571] env[62558]: INFO nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Terminating instance [ 531.228571] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquiring lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.228571] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquired lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.228571] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 531.367994] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.394417] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.394845] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.395364] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.395615] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.395878] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.396302] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.396902] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.397132] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.397353] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.397547] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.397748] env[62558]: DEBUG nova.virt.hardware [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.398670] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7711aa-8ed3-4eac-9779-30d5937fd9c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.408359] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cd40f0-f539-421c-8d8e-37df95c3662e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.456766] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Successfully created port: ce0a2b84-a054-4344-9de1-274fd798155b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.522934] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266477, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.523542] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 531.524212] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 531.524937] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 531.696385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.697012] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.703018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.631s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.703018] env[62558]: INFO nova.compute.claims [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.781196] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.012949] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.202666] env[62558]: DEBUG nova.compute.utils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.211031] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.211031] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 532.407844] env[62558]: DEBUG nova.policy [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8f49d29496540a883611670b2fafee4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa1d90fab22044528de22ba940b9a2a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.516267] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Releasing lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.516767] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.516970] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.517299] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d1b29f4-f535-4f50-95c4-0b81ebbd4153 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.529353] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bd9496-369d-4813-aca2-ed9ec1577e67 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.557595] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquiring lock "f52921c6-abbc-4a27-a451-e1cb4f5dfc48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.558069] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "f52921c6-abbc-4a27-a451-e1cb4f5dfc48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.564956] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 158e242b-3f2d-4cac-b3f3-c51da7f78bb1 could not be found. [ 532.565724] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.566090] env[62558]: INFO nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 532.566424] env[62558]: DEBUG oslo.service.loopingcall [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.566658] env[62558]: DEBUG nova.compute.manager [-] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.566911] env[62558]: DEBUG nova.network.neutron [-] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.581560] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.581825] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.581982] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.583569] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.583569] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.583720] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.583899] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.584190] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.584259] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.584630] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.584630] env[62558]: DEBUG nova.virt.hardware [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.585728] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e370b02c-d167-4321-8556-9a3138fceb31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.599528] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e145daf-5b39-49ad-b94c-aa6bf65fb1c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.616868] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 532.622819] env[62558]: DEBUG oslo.service.loopingcall [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.623799] env[62558]: DEBUG nova.network.neutron [-] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.625074] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 532.625626] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8e110d1-5070-40ac-bb95-bb7c209e4db1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.644901] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 532.644901] env[62558]: value = "task-1266478" [ 532.644901] env[62558]: _type = "Task" [ 532.644901] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.653305] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266478, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.673781] env[62558]: ERROR nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 532.673781] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.673781] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.673781] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.673781] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.673781] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.673781] env[62558]: ERROR nova.compute.manager raise self.value [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.673781] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.673781] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.673781] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.674237] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.674237] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.674237] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 532.674237] env[62558]: ERROR nova.compute.manager [ 532.674237] env[62558]: Traceback (most recent call last): [ 532.674237] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.674237] env[62558]: listener.cb(fileno) [ 532.674237] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.674237] env[62558]: result = function(*args, **kwargs) [ 532.674237] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.674237] env[62558]: return func(*args, **kwargs) [ 532.674237] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.674237] env[62558]: raise e [ 532.674237] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.674237] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 532.674237] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.674237] env[62558]: created_port_ids = self._update_ports_for_instance( [ 532.674237] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.674237] env[62558]: with excutils.save_and_reraise_exception(): [ 532.674237] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.674237] env[62558]: self.force_reraise() [ 532.674237] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.674237] env[62558]: raise self.value [ 532.674237] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.674237] env[62558]: updated_port = self._update_port( [ 532.674237] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.674237] env[62558]: _ensure_no_port_binding_failure(port) [ 532.674237] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.674237] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.674976] env[62558]: nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 532.674976] env[62558]: Removing descriptor: 15 [ 532.675517] env[62558]: ERROR nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Traceback (most recent call last): [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] yield resources [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.driver.spawn(context, instance, image_meta, [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] vm_ref = self.build_virtual_machine(instance, [ 532.675517] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] for vif in network_info: [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self._sync_wrapper(fn, *args, **kwargs) [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.wait() [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self[:] = self._gt.wait() [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self._exit_event.wait() [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.675886] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] result = hub.switch() [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self.greenlet.switch() [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] result = function(*args, **kwargs) [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return func(*args, **kwargs) [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise e [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] nwinfo = self.network_api.allocate_for_instance( [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] created_port_ids = self._update_ports_for_instance( [ 532.676245] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] with excutils.save_and_reraise_exception(): [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.force_reraise() [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise self.value [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] updated_port = self._update_port( [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] _ensure_no_port_binding_failure(port) [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise exception.PortBindingFailed(port_id=port['id']) [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 532.676569] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] [ 532.676953] env[62558]: INFO nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Terminating instance [ 532.679261] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.679441] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquired lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.679605] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.711105] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.985276] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007028e7-2e4f-4d63-81ac-dc047d2d0141 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.996020] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6698cd6-9ccd-423c-9853-4857e9214372 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.035349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409dd067-aae2-45de-ada3-ceca77cbb9e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.044204] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9a466b-d036-4fa8-9b68-ebbad2bdaefa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.059144] env[62558]: DEBUG nova.compute.provider_tree [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.126916] env[62558]: DEBUG nova.network.neutron [-] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.155545] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266478, 'name': CreateVM_Task, 'duration_secs': 0.387645} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.155729] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 533.156646] env[62558]: DEBUG oslo_vmware.service [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9622c04-ae9d-44a0-8eea-eefb57e9d3d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.163385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.163594] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.163995] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 533.164422] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc80b46b-9851-4fa4-837b-8106cf53802c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.169105] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 533.169105] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ce0f61-a26d-6373-12b9-4e94013783c1" [ 533.169105] env[62558]: _type = "Task" [ 533.169105] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.178183] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ce0f61-a26d-6373-12b9-4e94013783c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.290361] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.562162] env[62558]: DEBUG nova.scheduler.client.report [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.593191] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Successfully created port: 4901996c-557b-4324-90fc-d0762fb23d46 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.630921] env[62558]: INFO nova.compute.manager [-] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Took 1.06 seconds to deallocate network for instance. [ 533.634148] env[62558]: DEBUG nova.compute.claims [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.635011] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.686520] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.686939] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 533.687336] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.687644] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.688060] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 533.688274] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c7c116c-00b4-45a5-b6d3-ddcfec5df69b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.706433] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 533.706725] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 533.707943] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ad7a1c-63f1-44a3-97cc-e6f140940dea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.713512] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.718664] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41118d9d-6b23-4902-893f-a6aab3ce86a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.724162] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 533.724162] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e6ef43-8e91-0c90-2287-b3715260b7d4" [ 533.724162] env[62558]: _type = "Task" [ 533.724162] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.730069] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.739999] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e6ef43-8e91-0c90-2287-b3715260b7d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.771728] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.771972] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.772141] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.772341] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.772511] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.773190] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.773769] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.774026] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.774241] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.774432] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.776097] env[62558]: DEBUG nova.virt.hardware [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.776383] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6f0cfe-b933-40fd-a327-b78f5b9d2f6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.787986] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c574af-f2c5-46c7-a518-e17a3509149d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.831181] env[62558]: DEBUG nova.compute.manager [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Received event network-changed-6d159dac-efa3-4c50-828c-a4fc8e7e7c8d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.831181] env[62558]: DEBUG nova.compute.manager [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Refreshing instance network info cache due to event network-changed-6d159dac-efa3-4c50-828c-a4fc8e7e7c8d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.831181] env[62558]: DEBUG oslo_concurrency.lockutils [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] Acquiring lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.831181] env[62558]: DEBUG oslo_concurrency.lockutils [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] Acquired lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.831181] env[62558]: DEBUG nova.network.neutron [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Refreshing network info cache for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.068452] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.069015] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.072232] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.501s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.216475] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Releasing lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.217204] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.217424] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.217828] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e04e97b-d636-41d1-8978-dbf35be710a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.232640] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c712e9c1-46cc-40f9-baef-fc96285f0fe5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.254887] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Preparing fetch location {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 534.255956] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating directory with path [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 534.255956] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd4d356d-8575-43fb-91f9-62169558a336 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.264467] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 99296a9b-151c-4b32-8e6a-ad0e28945aac could not be found. [ 534.264761] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.264843] env[62558]: INFO nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Took 0.05 seconds to destroy the instance on the hypervisor. [ 534.265173] env[62558]: DEBUG oslo.service.loopingcall [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.266148] env[62558]: DEBUG nova.compute.manager [-] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.266148] env[62558]: DEBUG nova.network.neutron [-] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.286866] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Created directory with path [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 534.287091] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Fetch image to [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 534.287260] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Downloading image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk on the data store datastore1 {{(pid=62558) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 534.288143] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b237c553-9312-41ea-ba87-dc928e6f02c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.299037] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f654f7-3008-42ad-b769-5a36a78b0ac6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.314238] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf518ff-5e42-4196-93fd-bd2068a049e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.355281] env[62558]: DEBUG nova.network.neutron [-] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.357481] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d43dd86-6ba0-4ab6-8290-4c9203cf1e9e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.364071] env[62558]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-6836aaab-15c2-4e20-8b5b-aa6dad56e92f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.392322] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Downloading image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to the data store datastore1 {{(pid=62558) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 534.394899] env[62558]: DEBUG nova.network.neutron [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.473800] env[62558]: DEBUG oslo_vmware.rw_handles [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62558) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 534.544941] env[62558]: DEBUG nova.network.neutron [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.583237] env[62558]: DEBUG nova.compute.utils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.588193] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.590037] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 534.717913] env[62558]: DEBUG nova.compute.manager [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Received event network-changed-a8a4a427-c551-4893-abfe-62502b097be8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.718166] env[62558]: DEBUG nova.compute.manager [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Refreshing instance network info cache due to event network-changed-a8a4a427-c551-4893-abfe-62502b097be8. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.718350] env[62558]: DEBUG oslo_concurrency.lockutils [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] Acquiring lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.718487] env[62558]: DEBUG oslo_concurrency.lockutils [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] Acquired lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.718643] env[62558]: DEBUG nova.network.neutron [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Refreshing network info cache for port a8a4a427-c551-4893-abfe-62502b097be8 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.766025] env[62558]: DEBUG nova.policy [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df55140a6b014628a8648ea098246ef1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b853f0f0ad974c98a09ac690116770ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.860876] env[62558]: DEBUG nova.network.neutron [-] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.869930] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3a3eb7-aa7f-4edb-8582-fbd51f2c7722 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.885917] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b17413-7ab9-4ffe-bf03-f1c16f18a64d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.922721] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61831d32-faa1-42ce-a625-cb9681a009ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.931135] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea5c318-13b6-46b7-a44a-f65e47f5cf9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.948229] env[62558]: DEBUG nova.compute.provider_tree [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.048788] env[62558]: DEBUG oslo_concurrency.lockutils [req-19bf1c90-4567-4e7f-9aa8-4cc8f7d985e8 req-20cfe9d2-86ea-4383-99d2-dc8f03c62e9c service nova] Releasing lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.093278] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.255620] env[62558]: DEBUG oslo_vmware.rw_handles [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Completed reading data from the image iterator. {{(pid=62558) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 535.255874] env[62558]: DEBUG oslo_vmware.rw_handles [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 535.276480] env[62558]: DEBUG nova.network.neutron [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.319772] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Downloaded image file data bd09177c-5fd7-495c-893b-5ec219b6b3de to vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk on the data store datastore1 {{(pid=62558) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 535.322947] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Caching image {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 535.323258] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copying Virtual Disk [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk to [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 535.323554] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc695494-7859-4dc5-9cd5-d3ff500a188d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.334727] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 535.334727] env[62558]: value = "task-1266479" [ 535.334727] env[62558]: _type = "Task" [ 535.334727] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.345508] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.364943] env[62558]: INFO nova.compute.manager [-] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Took 1.10 seconds to deallocate network for instance. [ 535.367998] env[62558]: DEBUG nova.compute.claims [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.368216] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.457793] env[62558]: DEBUG nova.scheduler.client.report [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.648160] env[62558]: DEBUG nova.network.neutron [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.808245] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Successfully created port: 2864207b-caee-484d-808e-9ddbe63de986 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.848270] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266479, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.966028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.966028] env[62558]: ERROR nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Traceback (most recent call last): [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.driver.spawn(context, instance, image_meta, [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.966028] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] vm_ref = self.build_virtual_machine(instance, [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] for vif in network_info: [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self._sync_wrapper(fn, *args, **kwargs) [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.wait() [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self[:] = self._gt.wait() [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self._exit_event.wait() [ 535.966370] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] result = hub.switch() [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return self.greenlet.switch() [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] result = function(*args, **kwargs) [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] return func(*args, **kwargs) [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise e [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] nwinfo = self.network_api.allocate_for_instance( [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 535.966684] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] created_port_ids = self._update_ports_for_instance( [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] with excutils.save_and_reraise_exception(): [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] self.force_reraise() [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise self.value [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] updated_port = self._update_port( [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] _ensure_no_port_binding_failure(port) [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.967011] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] raise exception.PortBindingFailed(port_id=port['id']) [ 535.967302] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] nova.exception.PortBindingFailed: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. [ 535.967302] env[62558]: ERROR nova.compute.manager [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] [ 535.967302] env[62558]: DEBUG nova.compute.utils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 535.969246] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.791s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.970658] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.970823] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 535.973786] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.072s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.981773] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f0bbeb-d4e9-479c-9f5f-d9c69772a604 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.987762] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Build of instance 9b5c5a62-d708-4201-be94-28cff078b2e8 was re-scheduled: Binding failed for port b080e1c5-09b1-467b-a528-5319a1ab4c75, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.988331] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.988664] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquiring lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.988759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Acquired lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.988858] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 535.996356] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c427515-7a76-48f2-96a8-75c2ef43caed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.019602] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83361372-e3ca-4fe4-bf36-6b2299f87488 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.030433] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42dcfef-29d6-4c4b-9175-9316b54c221f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.069625] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181480MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 536.069625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.113842] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.148675] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.148942] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.149075] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.149254] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.149401] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.149524] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.150953] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.150953] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.150953] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.150953] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.150953] env[62558]: DEBUG nova.virt.hardware [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.151126] env[62558]: DEBUG oslo_concurrency.lockutils [req-a52dcea5-53e3-4cd2-9bf3-10a8a6cc7484 req-cdbd00c1-f890-48a6-81ed-8fa0df350961 service nova] Releasing lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.151974] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0090c9d7-9f98-4fef-805b-a0c07489ed78 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.163871] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525c8789-ba12-4206-9bd0-4043af6c7894 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.352133] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658978} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.352133] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copied Virtual Disk [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk to [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 536.352133] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleting the datastore file [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de/tmp-sparse.vmdk {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 536.352133] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-323838be-a8da-48a0-b4cf-cc0e37014f91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.360033] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 536.360033] env[62558]: value = "task-1266480" [ 536.360033] env[62558]: _type = "Task" [ 536.360033] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.373259] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.496897] env[62558]: ERROR nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 536.496897] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.496897] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.496897] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.496897] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.496897] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.496897] env[62558]: ERROR nova.compute.manager raise self.value [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.496897] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.496897] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.496897] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.497393] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.497393] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.497393] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 536.497393] env[62558]: ERROR nova.compute.manager [ 536.497393] env[62558]: Traceback (most recent call last): [ 536.497393] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.497393] env[62558]: listener.cb(fileno) [ 536.497393] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.497393] env[62558]: result = function(*args, **kwargs) [ 536.497393] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.497393] env[62558]: return func(*args, **kwargs) [ 536.497393] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.497393] env[62558]: raise e [ 536.497393] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.497393] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 536.497393] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.497393] env[62558]: created_port_ids = self._update_ports_for_instance( [ 536.497393] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.497393] env[62558]: with excutils.save_and_reraise_exception(): [ 536.497393] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.497393] env[62558]: self.force_reraise() [ 536.497393] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.497393] env[62558]: raise self.value [ 536.497393] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.497393] env[62558]: updated_port = self._update_port( [ 536.497393] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.497393] env[62558]: _ensure_no_port_binding_failure(port) [ 536.497393] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.497393] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.498616] env[62558]: nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 536.498616] env[62558]: Removing descriptor: 19 [ 536.498616] env[62558]: ERROR nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Traceback (most recent call last): [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] yield resources [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.driver.spawn(context, instance, image_meta, [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.498616] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] vm_ref = self.build_virtual_machine(instance, [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] for vif in network_info: [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self._sync_wrapper(fn, *args, **kwargs) [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.wait() [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self[:] = self._gt.wait() [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self._exit_event.wait() [ 536.498975] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] result = hub.switch() [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self.greenlet.switch() [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] result = function(*args, **kwargs) [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return func(*args, **kwargs) [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise e [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] nwinfo = self.network_api.allocate_for_instance( [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.499341] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] created_port_ids = self._update_ports_for_instance( [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] with excutils.save_and_reraise_exception(): [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.force_reraise() [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise self.value [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] updated_port = self._update_port( [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] _ensure_no_port_binding_failure(port) [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.499680] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise exception.PortBindingFailed(port_id=port['id']) [ 536.500023] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 536.500023] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] [ 536.500023] env[62558]: INFO nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Terminating instance [ 536.503060] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.503335] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquired lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.503552] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 536.582491] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.710031] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f041f6cc-0024-497a-8a85-4db001a47a0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.720234] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82efb26e-f0b9-4a3a-85bc-f51ac49b0d86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.755828] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d0c3f1-1b16-4859-b9c1-65d628df744a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.765164] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc1350d-d908-4b56-9b36-096ff4807a8b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.781277] env[62558]: DEBUG nova.compute.provider_tree [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.876505] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024104} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.876998] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 536.877181] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Moving file from [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73/bd09177c-5fd7-495c-893b-5ec219b6b3de to [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de. {{(pid=62558) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 536.877504] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-5a7c71be-2ffe-4747-950b-959e1d8f5939 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.887671] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 536.887671] env[62558]: value = "task-1266481" [ 536.887671] env[62558]: _type = "Task" [ 536.887671] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.896326] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266481, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.994730] env[62558]: DEBUG nova.compute.manager [req-2803e355-9d1f-4472-9d85-af5235922ba4 req-497c2c79-a974-4e66-8fe3-5823e6cc4923 service nova] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Received event network-vif-deleted-6d159dac-efa3-4c50-828c-a4fc8e7e7c8d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.073921] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.078944] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.128331] env[62558]: ERROR nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 537.128331] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.128331] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.128331] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.128331] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.128331] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.128331] env[62558]: ERROR nova.compute.manager raise self.value [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.128331] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.128331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.128331] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.128786] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.128786] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.128786] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 537.128786] env[62558]: ERROR nova.compute.manager [ 537.128786] env[62558]: Traceback (most recent call last): [ 537.128786] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.128786] env[62558]: listener.cb(fileno) [ 537.128786] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.128786] env[62558]: result = function(*args, **kwargs) [ 537.128786] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.128786] env[62558]: return func(*args, **kwargs) [ 537.128786] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.128786] env[62558]: raise e [ 537.128786] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.128786] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 537.128786] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.128786] env[62558]: created_port_ids = self._update_ports_for_instance( [ 537.128786] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.128786] env[62558]: with excutils.save_and_reraise_exception(): [ 537.128786] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.128786] env[62558]: self.force_reraise() [ 537.128786] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.128786] env[62558]: raise self.value [ 537.128786] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.128786] env[62558]: updated_port = self._update_port( [ 537.128786] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.128786] env[62558]: _ensure_no_port_binding_failure(port) [ 537.128786] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.128786] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.129667] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 537.129667] env[62558]: Removing descriptor: 18 [ 537.129667] env[62558]: ERROR nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Traceback (most recent call last): [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] yield resources [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.driver.spawn(context, instance, image_meta, [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.129667] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] vm_ref = self.build_virtual_machine(instance, [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] for vif in network_info: [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self._sync_wrapper(fn, *args, **kwargs) [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.wait() [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self[:] = self._gt.wait() [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self._exit_event.wait() [ 537.130032] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] result = hub.switch() [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self.greenlet.switch() [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] result = function(*args, **kwargs) [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return func(*args, **kwargs) [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise e [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] nwinfo = self.network_api.allocate_for_instance( [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.130400] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] created_port_ids = self._update_ports_for_instance( [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] with excutils.save_and_reraise_exception(): [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.force_reraise() [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise self.value [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] updated_port = self._update_port( [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] _ensure_no_port_binding_failure(port) [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.130792] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise exception.PortBindingFailed(port_id=port['id']) [ 537.131155] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 537.131155] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] [ 537.131155] env[62558]: INFO nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Terminating instance [ 537.134582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.134582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquired lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.134582] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.284372] env[62558]: DEBUG nova.scheduler.client.report [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.337786] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.400757] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266481, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026188} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.400757] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] File moved {{(pid=62558) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 537.401097] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Cleaning up location [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73 {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 537.401097] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleting the datastore file [datastore1] vmware_temp/a48f312c-af73-471f-82b8-2fb2c5796b73 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 537.401887] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c77cc18c-4349-4674-97a7-2677ca42a107 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.408676] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 537.408676] env[62558]: value = "task-1266482" [ 537.408676] env[62558]: _type = "Task" [ 537.408676] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.420312] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.529138] env[62558]: DEBUG nova.compute.manager [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Received event network-vif-deleted-a8a4a427-c551-4893-abfe-62502b097be8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.529348] env[62558]: DEBUG nova.compute.manager [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Received event network-changed-ce0a2b84-a054-4344-9de1-274fd798155b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.529515] env[62558]: DEBUG nova.compute.manager [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Refreshing instance network info cache due to event network-changed-ce0a2b84-a054-4344-9de1-274fd798155b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 537.529696] env[62558]: DEBUG oslo_concurrency.lockutils [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] Acquiring lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.582755] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Releasing lock "refresh_cache-9b5c5a62-d708-4201-be94-28cff078b2e8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.583397] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 537.583397] env[62558]: DEBUG nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.583540] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.601238] env[62558]: DEBUG nova.compute.manager [None req-720d8a51-39c2-4962-a739-3abd3fddd5a7 tempest-ServerDiagnosticsV248Test-857757544 tempest-ServerDiagnosticsV248Test-857757544-project-admin] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.601906] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574b02ce-5506-4740-ab75-8b0b8fb42f74 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.612506] env[62558]: INFO nova.compute.manager [None req-720d8a51-39c2-4962-a739-3abd3fddd5a7 tempest-ServerDiagnosticsV248Test-857757544 tempest-ServerDiagnosticsV248Test-857757544-project-admin] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Retrieving diagnostics [ 537.612506] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb27f873-f746-4958-9177-d1d4832b84aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.649602] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.679912] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.794380] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.795950] env[62558]: ERROR nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Traceback (most recent call last): [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.driver.spawn(context, instance, image_meta, [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] vm_ref = self.build_virtual_machine(instance, [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.795950] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] for vif in network_info: [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self._sync_wrapper(fn, *args, **kwargs) [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.wait() [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self[:] = self._gt.wait() [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self._exit_event.wait() [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] result = hub.switch() [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.796356] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return self.greenlet.switch() [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] result = function(*args, **kwargs) [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] return func(*args, **kwargs) [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise e [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] nwinfo = self.network_api.allocate_for_instance( [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] created_port_ids = self._update_ports_for_instance( [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] with excutils.save_and_reraise_exception(): [ 537.796675] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] self.force_reraise() [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise self.value [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] updated_port = self._update_port( [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] _ensure_no_port_binding_failure(port) [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] raise exception.PortBindingFailed(port_id=port['id']) [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] nova.exception.PortBindingFailed: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. [ 537.797039] env[62558]: ERROR nova.compute.manager [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] [ 537.797354] env[62558]: DEBUG nova.compute.utils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 537.797764] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.895s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.800036] env[62558]: INFO nova.compute.claims [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.804676] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Build of instance d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d was re-scheduled: Binding failed for port 90d1f7f3-3524-4853-a84d-8f3ed8fbce17, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 537.804676] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 537.804867] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquiring lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.804867] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Acquired lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.805390] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.845126] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Releasing lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.845126] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 537.845126] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 537.845126] env[62558]: DEBUG oslo_concurrency.lockutils [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] Acquired lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.845126] env[62558]: DEBUG nova.network.neutron [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Refreshing network info cache for port ce0a2b84-a054-4344-9de1-274fd798155b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 537.846282] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbf4b060-ec76-489f-81db-b3d2dfdadb6d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.859036] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b0d773-c7d2-43e0-a705-52cd8a0a2748 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.891644] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f923c30-6c68-4e5d-be8a-4f18e996a999 could not be found. [ 537.892469] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 537.895139] env[62558]: INFO nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Took 0.05 seconds to destroy the instance on the hypervisor. [ 537.895139] env[62558]: DEBUG oslo.service.loopingcall [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.895139] env[62558]: DEBUG nova.compute.manager [-] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.895139] env[62558]: DEBUG nova.network.neutron [-] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.921363] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024299} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.921653] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 537.922431] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73c31520-35ec-4dc4-92f0-5a6b57243dec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.928573] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 537.928573] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52477544-4d55-6d42-478d-d6961cc28b35" [ 537.928573] env[62558]: _type = "Task" [ 537.928573] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.939604] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52477544-4d55-6d42-478d-d6961cc28b35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.973808] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.057305] env[62558]: DEBUG nova.network.neutron [-] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.153428] env[62558]: DEBUG nova.network.neutron [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.342504] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.396566] env[62558]: DEBUG nova.network.neutron [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.439809] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52477544-4d55-6d42-478d-d6961cc28b35, 'name': SearchDatastore_Task, 'duration_secs': 0.009438} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.440101] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.440349] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 538.440616] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdb869e8-f607-4e49-bfbd-e1b17a1afb89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.448085] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 538.448085] env[62558]: value = "task-1266483" [ 538.448085] env[62558]: _type = "Task" [ 538.448085] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.459155] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266483, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.460520] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.478625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Releasing lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.479139] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.479365] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.479639] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e795a49e-90b7-4723-80ae-d1db3b34975f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.490431] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb460602-32ee-4e71-91ef-d4dbdf500268 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.518096] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d66f026-8b3e-4670-ada6-0109d99bb984 could not be found. [ 538.518096] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.518096] env[62558]: INFO nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.518096] env[62558]: DEBUG oslo.service.loopingcall [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.519268] env[62558]: DEBUG nova.network.neutron [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.522944] env[62558]: DEBUG nova.compute.manager [-] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.522944] env[62558]: DEBUG nova.network.neutron [-] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.547873] env[62558]: DEBUG nova.network.neutron [-] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.562168] env[62558]: DEBUG nova.network.neutron [-] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.664022] env[62558]: INFO nova.compute.manager [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] Took 1.08 seconds to deallocate network for instance. [ 538.687698] env[62558]: ERROR nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 538.687698] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.687698] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.687698] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.687698] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.687698] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.687698] env[62558]: ERROR nova.compute.manager raise self.value [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.687698] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.687698] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.687698] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.688151] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.688151] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.688151] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 538.688151] env[62558]: ERROR nova.compute.manager [ 538.688151] env[62558]: Traceback (most recent call last): [ 538.688151] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.688151] env[62558]: listener.cb(fileno) [ 538.688151] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.688151] env[62558]: result = function(*args, **kwargs) [ 538.688151] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.688151] env[62558]: return func(*args, **kwargs) [ 538.688151] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.688151] env[62558]: raise e [ 538.688151] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.688151] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 538.688151] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.688151] env[62558]: created_port_ids = self._update_ports_for_instance( [ 538.688151] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.688151] env[62558]: with excutils.save_and_reraise_exception(): [ 538.688151] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.688151] env[62558]: self.force_reraise() [ 538.688151] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.688151] env[62558]: raise self.value [ 538.688151] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.688151] env[62558]: updated_port = self._update_port( [ 538.688151] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.688151] env[62558]: _ensure_no_port_binding_failure(port) [ 538.688151] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.688151] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.688964] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 538.688964] env[62558]: Removing descriptor: 22 [ 538.688964] env[62558]: ERROR nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Traceback (most recent call last): [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] yield resources [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.driver.spawn(context, instance, image_meta, [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.688964] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] vm_ref = self.build_virtual_machine(instance, [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] for vif in network_info: [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self._sync_wrapper(fn, *args, **kwargs) [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.wait() [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self[:] = self._gt.wait() [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self._exit_event.wait() [ 538.689315] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] result = hub.switch() [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self.greenlet.switch() [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] result = function(*args, **kwargs) [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return func(*args, **kwargs) [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise e [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] nwinfo = self.network_api.allocate_for_instance( [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.689643] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] created_port_ids = self._update_ports_for_instance( [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] with excutils.save_and_reraise_exception(): [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.force_reraise() [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise self.value [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] updated_port = self._update_port( [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] _ensure_no_port_binding_failure(port) [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.689954] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise exception.PortBindingFailed(port_id=port['id']) [ 538.690276] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 538.690276] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] [ 538.690276] env[62558]: INFO nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Terminating instance [ 538.692514] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquiring lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.692797] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquired lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.693646] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.964608] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Releasing lock "refresh_cache-d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.964848] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 538.965019] env[62558]: DEBUG nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.965185] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.971415] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266483, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.989498] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.023643] env[62558]: DEBUG oslo_concurrency.lockutils [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] Releasing lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.023643] env[62558]: DEBUG nova.compute.manager [req-3937c2d0-94b7-4cce-a702-3a2aec4ae8f2 req-9984cb4c-148d-436f-9491-a5308e94d663 service nova] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Received event network-vif-deleted-ce0a2b84-a054-4344-9de1-274fd798155b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.050918] env[62558]: DEBUG nova.network.neutron [-] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.069362] env[62558]: INFO nova.compute.manager [-] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Took 1.18 seconds to deallocate network for instance. [ 539.076362] env[62558]: DEBUG nova.compute.claims [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.076362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.090631] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baed03eb-60a2-4389-913f-1d9919b253f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.101782] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a85eade-8f7c-498e-908e-e29ff256d95a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.135553] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebb8a4b-956d-4829-aba3-ce0ad20067d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.143569] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70049a0-2305-42d5-8f58-29a0691df5a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.157063] env[62558]: DEBUG nova.compute.provider_tree [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.224837] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.349426] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.461302] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266483, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58047} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.463082] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 539.463987] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 539.464528] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7b139b1-ccc8-4b62-8bcc-463d832dc2a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.471855] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 539.471855] env[62558]: value = "task-1266484" [ 539.471855] env[62558]: _type = "Task" [ 539.471855] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.480100] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266484, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.492093] env[62558]: DEBUG nova.network.neutron [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.556175] env[62558]: INFO nova.compute.manager [-] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Took 1.03 seconds to deallocate network for instance. [ 539.556832] env[62558]: DEBUG nova.compute.claims [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.557164] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.659706] env[62558]: DEBUG nova.scheduler.client.report [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.701804] env[62558]: INFO nova.scheduler.client.report [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Deleted allocations for instance 9b5c5a62-d708-4201-be94-28cff078b2e8 [ 539.811558] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.811846] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.812086] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.812979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.812979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.815372] env[62558]: INFO nova.compute.manager [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Terminating instance [ 539.819743] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "refresh_cache-06df57a7-2c53-4f79-bec4-e46cfe1dca63" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.819926] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquired lock "refresh_cache-06df57a7-2c53-4f79-bec4-e46cfe1dca63" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.820203] env[62558]: DEBUG nova.network.neutron [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.853034] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Releasing lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.853034] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.853258] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.853905] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-603fa30c-eeec-41ee-9a87-091c32f3cb93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.870934] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b69f2c-ecda-460b-b1c6-467f59fb4751 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.898183] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b00d6153-c94f-4605-9b97-9a3a813ee93b could not be found. [ 539.898302] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.898582] env[62558]: INFO nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 539.898920] env[62558]: DEBUG oslo.service.loopingcall [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.899178] env[62558]: DEBUG nova.compute.manager [-] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.899343] env[62558]: DEBUG nova.network.neutron [-] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.925880] env[62558]: DEBUG nova.network.neutron [-] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.985012] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266484, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108868} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.985493] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 539.986505] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ceb0483-9a23-4ddc-ac38-4ebe8e30421d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.002572] env[62558]: INFO nova.compute.manager [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] [instance: d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d] Took 1.04 seconds to deallocate network for instance. [ 540.013338] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 540.014357] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b265f2e0-bd1f-4e8d-b881-291beec2821c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.036663] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 540.036663] env[62558]: value = "task-1266485" [ 540.036663] env[62558]: _type = "Task" [ 540.036663] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.045153] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.110755] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Received event network-changed-4901996c-557b-4324-90fc-d0762fb23d46 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.110970] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Refreshing instance network info cache due to event network-changed-4901996c-557b-4324-90fc-d0762fb23d46. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 540.111194] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Acquiring lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.111330] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Acquired lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.111919] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Refreshing network info cache for port 4901996c-557b-4324-90fc-d0762fb23d46 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 540.164875] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.169096] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.176198] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.671s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.176408] env[62558]: INFO nova.compute.claims [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.213133] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7e56b079-5300-4878-8958-f421775c818d tempest-FloatingIPsAssociationTestJSON-905131595 tempest-FloatingIPsAssociationTestJSON-905131595-project-member] Lock "9b5c5a62-d708-4201-be94-28cff078b2e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.653s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.217221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "9b5c5a62-d708-4201-be94-28cff078b2e8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.464s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.217221] env[62558]: INFO nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9b5c5a62-d708-4201-be94-28cff078b2e8] During sync_power_state the instance has a pending task (spawning). Skip. [ 540.217221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "9b5c5a62-d708-4201-be94-28cff078b2e8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.429550] env[62558]: DEBUG nova.network.neutron [-] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.450963] env[62558]: DEBUG nova.network.neutron [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.548850] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266485, 'name': ReconfigVM_Task, 'duration_secs': 0.291519} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.549219] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfigured VM instance instance-00000003 to attach disk [datastore1] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 540.549957] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4606f6da-6118-4332-b239-89e10a54ea53 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.561821] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 540.561821] env[62558]: value = "task-1266486" [ 540.561821] env[62558]: _type = "Task" [ 540.561821] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.576853] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266486, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.627340] env[62558]: DEBUG nova.network.neutron [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.657029] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.687859] env[62558]: DEBUG nova.compute.utils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.691040] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.691040] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 540.719958] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.854603] env[62558]: DEBUG nova.policy [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '981bde3f516747fab7dccfc98e102729', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e491249e28d74357a6f22a8c31c5a32e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.908917] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.931612] env[62558]: INFO nova.compute.manager [-] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Took 1.03 seconds to deallocate network for instance. [ 540.945828] env[62558]: DEBUG nova.compute.claims [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.946196] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.053947] env[62558]: INFO nova.scheduler.client.report [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Deleted allocations for instance d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d [ 541.075801] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266486, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.130405] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Releasing lock "refresh_cache-06df57a7-2c53-4f79-bec4-e46cfe1dca63" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.132698] env[62558]: DEBUG nova.compute.manager [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.133115] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.135114] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ce521e-c5ec-49bc-8c21-738e5d2e3c6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.144159] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 541.144159] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-725abcb3-735f-4565-9e07-95f3c9e62586 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.150798] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 541.150798] env[62558]: value = "task-1266487" [ 541.150798] env[62558]: _type = "Task" [ 541.150798] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.161083] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.190851] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.250769] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.409691] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Successfully created port: 54187e76-94e3-4303-be6d-5241a92560bb {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.412937] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Releasing lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.414820] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Received event network-vif-deleted-4901996c-557b-4324-90fc-d0762fb23d46 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.414820] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Received event network-changed-2864207b-caee-484d-808e-9ddbe63de986 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.414820] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Refreshing instance network info cache due to event network-changed-2864207b-caee-484d-808e-9ddbe63de986. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.414820] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Acquiring lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.414820] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Acquired lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.415011] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Refreshing network info cache for port 2864207b-caee-484d-808e-9ddbe63de986 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 541.427212] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614ad9b9-e8bf-425e-8ac8-7eb990ae1980 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.436601] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084455ed-5020-4727-8da8-d7e09f92b400 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.478388] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adf9402-1cbd-448f-959b-4aa3122a940b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.486646] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa55535-c81c-4d3f-848f-e6b00c1b8c3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.507624] env[62558]: DEBUG nova.compute.provider_tree [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.523221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquiring lock "5fc9a508-323d-4f1c-ae46-d15f340f6b73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.524065] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "5fc9a508-323d-4f1c-ae46-d15f340f6b73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.569632] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc17e1b0-ad76-4906-871d-ea524c4761a7 tempest-ServersAdminNegativeTestJSON-1557886903 tempest-ServersAdminNegativeTestJSON-1557886903-project-member] Lock "d8a5ccf1-34dd-42b2-aa83-3d4cdc2df55d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.297s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.577209] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266486, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.668264] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266487, 'name': PowerOffVM_Task, 'duration_secs': 0.12267} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.668515] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 541.668676] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 541.668920] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d38834c9-cb1c-4843-88f9-dbd8f34e5c31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.694013] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 541.694909] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 541.694909] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleting the datastore file [datastore2] 06df57a7-2c53-4f79-bec4-e46cfe1dca63 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 541.694909] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e01b606a-2643-40f0-92a9-8bfd4606cf0c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.706906] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for the task: (returnval){ [ 541.706906] env[62558]: value = "task-1266489" [ 541.706906] env[62558]: _type = "Task" [ 541.706906] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.716840] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.937617] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.011585] env[62558]: DEBUG nova.scheduler.client.report [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.059349] env[62558]: DEBUG nova.network.neutron [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.077376] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266486, 'name': Rename_Task, 'duration_secs': 1.154683} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.077796] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.081820] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 542.086517] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd57f5a5-92ab-41e5-b40a-900cae231cd0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.094964] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 542.094964] env[62558]: value = "task-1266490" [ 542.094964] env[62558]: _type = "Task" [ 542.094964] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.105728] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.172895] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquiring lock "5845d273-b18a-447a-ad2d-a011bfcc2423" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.173143] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "5845d273-b18a-447a-ad2d-a011bfcc2423" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.202684] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.216797] env[62558]: DEBUG oslo_vmware.api [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Task: {'id': task-1266489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089322} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.219945] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 542.220221] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 542.220610] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.220610] env[62558]: INFO nova.compute.manager [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Took 1.09 seconds to destroy the instance on the hypervisor. [ 542.220700] env[62558]: DEBUG oslo.service.loopingcall [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.222512] env[62558]: DEBUG nova.compute.manager [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.222512] env[62558]: DEBUG nova.network.neutron [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.228498] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.228863] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.228863] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.229065] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.229212] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.229357] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.229587] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.229710] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.229899] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.230398] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.231102] env[62558]: DEBUG nova.virt.hardware [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.232349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4457cada-5612-4501-b1a7-903d38fbb182 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.241619] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d8332d-54ff-4a4f-8b61-6589cb0999ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.261025] env[62558]: DEBUG nova.network.neutron [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.516862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.518933] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.523893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.918s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.562515] env[62558]: DEBUG oslo_concurrency.lockutils [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] Releasing lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.565719] env[62558]: DEBUG nova.compute.manager [req-08150546-e754-4ce3-9618-598767d8323a req-385e4118-ebd9-48c3-8b5b-5cd7ca23aa72 service nova] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Received event network-vif-deleted-2864207b-caee-484d-808e-9ddbe63de986 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.608956] env[62558]: DEBUG oslo_vmware.api [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266490, 'name': PowerOnVM_Task, 'duration_secs': 0.471196} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.609289] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 542.609932] env[62558]: DEBUG nova.compute.manager [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 542.610255] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931eefe4-47f5-46de-b3ee-6271f9b0dfa8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.614908] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.670365] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.670546] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.766789] env[62558]: DEBUG nova.network.neutron [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.034617] env[62558]: DEBUG nova.compute.utils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.034617] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.034617] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 543.135076] env[62558]: DEBUG nova.policy [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.149629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.273181] env[62558]: INFO nova.compute.manager [-] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Took 1.05 seconds to deallocate network for instance. [ 543.293859] env[62558]: ERROR nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 543.293859] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.293859] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.293859] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.293859] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.293859] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.293859] env[62558]: ERROR nova.compute.manager raise self.value [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.293859] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.293859] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.293859] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.294331] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.294331] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.294331] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 543.294331] env[62558]: ERROR nova.compute.manager [ 543.294331] env[62558]: Traceback (most recent call last): [ 543.294331] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.294331] env[62558]: listener.cb(fileno) [ 543.294331] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.294331] env[62558]: result = function(*args, **kwargs) [ 543.294331] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.294331] env[62558]: return func(*args, **kwargs) [ 543.294331] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.294331] env[62558]: raise e [ 543.294331] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.294331] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 543.294331] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.294331] env[62558]: created_port_ids = self._update_ports_for_instance( [ 543.294331] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.294331] env[62558]: with excutils.save_and_reraise_exception(): [ 543.294331] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.294331] env[62558]: self.force_reraise() [ 543.294331] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.294331] env[62558]: raise self.value [ 543.294331] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.294331] env[62558]: updated_port = self._update_port( [ 543.294331] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.294331] env[62558]: _ensure_no_port_binding_failure(port) [ 543.294331] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.294331] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.295107] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 543.295107] env[62558]: Removing descriptor: 22 [ 543.295107] env[62558]: ERROR nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Traceback (most recent call last): [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] yield resources [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.driver.spawn(context, instance, image_meta, [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.295107] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] vm_ref = self.build_virtual_machine(instance, [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] for vif in network_info: [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self._sync_wrapper(fn, *args, **kwargs) [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.wait() [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self[:] = self._gt.wait() [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self._exit_event.wait() [ 543.295437] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] result = hub.switch() [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self.greenlet.switch() [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] result = function(*args, **kwargs) [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return func(*args, **kwargs) [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise e [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] nwinfo = self.network_api.allocate_for_instance( [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.296135] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] created_port_ids = self._update_ports_for_instance( [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] with excutils.save_and_reraise_exception(): [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.force_reraise() [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise self.value [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] updated_port = self._update_port( [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] _ensure_no_port_binding_failure(port) [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.297099] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise exception.PortBindingFailed(port_id=port['id']) [ 543.297393] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 543.297393] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] [ 543.297393] env[62558]: INFO nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Terminating instance [ 543.297393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquiring lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.297393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquired lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.297393] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.370929] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "f4f7e2c9-229c-4547-b674-5a5219a28151" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.371235] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.391921] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218a9df6-5e09-464e-9f03-6ab1cd26dbc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.400012] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a1f083-88ce-4c52-ba2e-4d11f93ec3c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.436548] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquiring lock "813e287c-9d7f-4e9c-9216-986f631b6eb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.437655] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "813e287c-9d7f-4e9c-9216-986f631b6eb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.440124] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2f1821-4a28-4ea2-90cd-e9630dccc7a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.448761] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a7de0c-f14d-4329-b139-757ce51af192 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.463491] env[62558]: DEBUG nova.compute.provider_tree [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.546079] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.650380] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Successfully created port: 6996c67c-5f06-4713-aced-7d64fd17e0f1 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.785302] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.812545] env[62558]: DEBUG nova.compute.manager [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Received event network-changed-54187e76-94e3-4303-be6d-5241a92560bb {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.812742] env[62558]: DEBUG nova.compute.manager [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Refreshing instance network info cache due to event network-changed-54187e76-94e3-4303-be6d-5241a92560bb. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.812935] env[62558]: DEBUG oslo_concurrency.lockutils [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] Acquiring lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.820407] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.968837] env[62558]: DEBUG nova.scheduler.client.report [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.022112] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.479596] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.479596] env[62558]: ERROR nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Traceback (most recent call last): [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.driver.spawn(context, instance, image_meta, [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.479596] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] vm_ref = self.build_virtual_machine(instance, [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] for vif in network_info: [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self._sync_wrapper(fn, *args, **kwargs) [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.wait() [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self[:] = self._gt.wait() [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self._exit_event.wait() [ 544.479878] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] result = hub.switch() [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return self.greenlet.switch() [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] result = function(*args, **kwargs) [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] return func(*args, **kwargs) [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise e [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] nwinfo = self.network_api.allocate_for_instance( [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.480235] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] created_port_ids = self._update_ports_for_instance( [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] with excutils.save_and_reraise_exception(): [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] self.force_reraise() [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise self.value [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] updated_port = self._update_port( [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] _ensure_no_port_binding_failure(port) [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.480557] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] raise exception.PortBindingFailed(port_id=port['id']) [ 544.480877] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] nova.exception.PortBindingFailed: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. [ 544.480877] env[62558]: ERROR nova.compute.manager [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] [ 544.480877] env[62558]: DEBUG nova.compute.utils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.480877] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.846s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.484537] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Build of instance d6eaf42d-dd83-4158-a507-9bc71cd1aa43 was re-scheduled: Binding failed for port 346a4760-dcb6-456b-8249-a1c7fa4c00a7, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.485068] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.485437] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.485646] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquired lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.485848] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.528775] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Releasing lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.529679] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.529679] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.530160] env[62558]: DEBUG oslo_concurrency.lockutils [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] Acquired lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.530386] env[62558]: DEBUG nova.network.neutron [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Refreshing network info cache for port 54187e76-94e3-4303-be6d-5241a92560bb {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.532151] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c368777a-da69-4c81-a74f-9dcf50b63667 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.542155] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57466b8e-60a8-4d05-8779-5bedc3a95b90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.559960] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.576255] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c173701-9f61-4561-ac3c-7ac3de4101d9 could not be found. [ 544.576255] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.576255] env[62558]: INFO nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.576696] env[62558]: DEBUG oslo.service.loopingcall [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.576696] env[62558]: DEBUG nova.compute.manager [-] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.576696] env[62558]: DEBUG nova.network.neutron [-] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.595931] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.596126] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.596200] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.596475] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.596705] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.596940] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.598141] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.598141] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.598141] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.598141] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.598141] env[62558]: DEBUG nova.virt.hardware [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.599446] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a043e5c-55a2-4017-b909-67581a86d8dd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.612736] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ed6491-d41f-4ecd-be14-d6cf4123180d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.627948] env[62558]: DEBUG nova.network.neutron [-] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.064677] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.069561] env[62558]: DEBUG nova.network.neutron [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.130955] env[62558]: DEBUG nova.network.neutron [-] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.193568] env[62558]: DEBUG nova.network.neutron [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.279587] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.299334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6c2702-585f-4126-aeb9-fdeffd2ae65e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.307740] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b23bc4-3266-4e7b-93e3-bcd099cdcfea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.343768] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88570bfd-f344-41fc-8ef5-41c26b7b8cf8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.351543] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c3b834-fa1a-4103-9e75-b19efd33135f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.365885] env[62558]: DEBUG nova.compute.provider_tree [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.509382] env[62558]: INFO nova.compute.manager [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Rebuilding instance [ 545.571213] env[62558]: DEBUG nova.compute.manager [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.572097] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622b46cc-f69e-4ccc-98e3-19de02d3ae69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.633607] env[62558]: INFO nova.compute.manager [-] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Took 1.06 seconds to deallocate network for instance. [ 545.638811] env[62558]: DEBUG nova.compute.claims [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.639015] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.698627] env[62558]: DEBUG oslo_concurrency.lockutils [req-24fbab31-e8c8-49e7-9d5c-4260c3cd85c2 req-b6be3600-4f1e-4cf0-941c-acf2d07fd454 service nova] Releasing lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.783700] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Releasing lock "refresh_cache-d6eaf42d-dd83-4158-a507-9bc71cd1aa43" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.784050] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.784364] env[62558]: DEBUG nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.784491] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.847415] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.869592] env[62558]: DEBUG nova.scheduler.client.report [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.083462] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 546.083837] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-193cce1d-2574-44bb-afed-a340608cc233 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.092233] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 546.092233] env[62558]: value = "task-1266491" [ 546.092233] env[62558]: _type = "Task" [ 546.092233] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.101098] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.350405] env[62558]: DEBUG nova.network.neutron [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.374546] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.375205] env[62558]: ERROR nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Traceback (most recent call last): [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.driver.spawn(context, instance, image_meta, [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] vm_ref = self.build_virtual_machine(instance, [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.375205] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] for vif in network_info: [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self._sync_wrapper(fn, *args, **kwargs) [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.wait() [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self[:] = self._gt.wait() [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self._exit_event.wait() [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] result = hub.switch() [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.375580] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return self.greenlet.switch() [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] result = function(*args, **kwargs) [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] return func(*args, **kwargs) [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise e [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] nwinfo = self.network_api.allocate_for_instance( [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] created_port_ids = self._update_ports_for_instance( [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] with excutils.save_and_reraise_exception(): [ 546.376036] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] self.force_reraise() [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise self.value [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] updated_port = self._update_port( [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] _ensure_no_port_binding_failure(port) [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] raise exception.PortBindingFailed(port_id=port['id']) [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] nova.exception.PortBindingFailed: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. [ 546.377096] env[62558]: ERROR nova.compute.manager [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] [ 546.377385] env[62558]: DEBUG nova.compute.utils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 546.380735] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.009s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.383833] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Build of instance 158e242b-3f2d-4cac-b3f3-c51da7f78bb1 was re-scheduled: Binding failed for port 6d159dac-efa3-4c50-828c-a4fc8e7e7c8d, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 546.384360] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 546.385987] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquiring lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.386152] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Acquired lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.386354] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.602556] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266491, 'name': PowerOffVM_Task, 'duration_secs': 0.11946} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.602803] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 546.603205] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 546.604775] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516a11fc-8871-4e08-af21-aa3f2e7f7e61 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.611978] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 546.611978] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5345aec2-4fd4-4997-9f3d-163c90980ef1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.647030] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 546.647198] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 546.647370] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Deleting the datastore file [datastore1] d02e6383-42d9-40db-89a3-ba8b5595b95b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.647628] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca05e002-68eb-4e59-9680-d6f7717d460d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.654355] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 546.654355] env[62558]: value = "task-1266493" [ 546.654355] env[62558]: _type = "Task" [ 546.654355] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.662713] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.700477] env[62558]: ERROR nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 546.700477] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.700477] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.700477] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.700477] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.700477] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.700477] env[62558]: ERROR nova.compute.manager raise self.value [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.700477] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.700477] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.700477] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.701039] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.701039] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.701039] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 546.701039] env[62558]: ERROR nova.compute.manager [ 546.701039] env[62558]: Traceback (most recent call last): [ 546.701039] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.701039] env[62558]: listener.cb(fileno) [ 546.701039] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.701039] env[62558]: result = function(*args, **kwargs) [ 546.701039] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.701039] env[62558]: return func(*args, **kwargs) [ 546.701039] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.701039] env[62558]: raise e [ 546.701039] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.701039] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 546.701039] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.701039] env[62558]: created_port_ids = self._update_ports_for_instance( [ 546.701039] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.701039] env[62558]: with excutils.save_and_reraise_exception(): [ 546.701039] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.701039] env[62558]: self.force_reraise() [ 546.701039] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.701039] env[62558]: raise self.value [ 546.701039] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.701039] env[62558]: updated_port = self._update_port( [ 546.701039] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.701039] env[62558]: _ensure_no_port_binding_failure(port) [ 546.701039] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.701039] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.701862] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 546.701862] env[62558]: Removing descriptor: 18 [ 546.701862] env[62558]: ERROR nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Traceback (most recent call last): [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] yield resources [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.driver.spawn(context, instance, image_meta, [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.701862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] vm_ref = self.build_virtual_machine(instance, [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] for vif in network_info: [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self._sync_wrapper(fn, *args, **kwargs) [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.wait() [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self[:] = self._gt.wait() [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self._exit_event.wait() [ 546.702189] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] result = hub.switch() [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self.greenlet.switch() [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] result = function(*args, **kwargs) [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return func(*args, **kwargs) [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise e [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] nwinfo = self.network_api.allocate_for_instance( [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.702516] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] created_port_ids = self._update_ports_for_instance( [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] with excutils.save_and_reraise_exception(): [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.force_reraise() [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise self.value [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] updated_port = self._update_port( [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] _ensure_no_port_binding_failure(port) [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.702893] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise exception.PortBindingFailed(port_id=port['id']) [ 546.703219] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 546.703219] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] [ 546.703219] env[62558]: INFO nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Terminating instance [ 546.704369] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.704526] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.704686] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.854018] env[62558]: INFO nova.compute.manager [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: d6eaf42d-dd83-4158-a507-9bc71cd1aa43] Took 1.07 seconds to deallocate network for instance. [ 546.953654] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.141359] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.174954] env[62558]: DEBUG nova.compute.manager [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Received event network-vif-deleted-54187e76-94e3-4303-be6d-5241a92560bb {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.175173] env[62558]: DEBUG nova.compute.manager [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Received event network-changed-6996c67c-5f06-4713-aced-7d64fd17e0f1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.175328] env[62558]: DEBUG nova.compute.manager [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Refreshing instance network info cache due to event network-changed-6996c67c-5f06-4713-aced-7d64fd17e0f1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 547.175543] env[62558]: DEBUG oslo_concurrency.lockutils [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] Acquiring lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.176242] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086957} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.176601] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.176684] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 547.176779] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.203526] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f9205e-c521-45e4-9af9-08087ba21235 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.214979] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec336bac-3a6d-4026-8d8b-929016aa663d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.249656] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.252354] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c7ef2a-69e4-4277-9e42-7fcd0ab9a932 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.261015] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06c52a7-4fd7-4dd8-b434-b113296e8064 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.276235] env[62558]: DEBUG nova.compute.provider_tree [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.494192] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.496898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.496898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.645853] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Releasing lock "refresh_cache-158e242b-3f2d-4cac-b3f3-c51da7f78bb1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.646152] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 547.646319] env[62558]: DEBUG nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.646483] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.682884] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.782019] env[62558]: DEBUG nova.scheduler.client.report [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.913483] env[62558]: INFO nova.scheduler.client.report [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Deleted allocations for instance d6eaf42d-dd83-4158-a507-9bc71cd1aa43 [ 547.998758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.998758] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.998840] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.999121] env[62558]: DEBUG oslo_concurrency.lockutils [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] Acquired lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.999290] env[62558]: DEBUG nova.network.neutron [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Refreshing network info cache for port 6996c67c-5f06-4713-aced-7d64fd17e0f1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 548.000413] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e74a0a3-98eb-40f6-8130-3977a30eaeab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.012873] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d99e0d-816d-41b1-823e-02f3ba9e1564 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.039578] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b08cb5a2-56d1-43da-ae4a-684077dd517f could not be found. [ 548.039578] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.039578] env[62558]: INFO nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 548.039578] env[62558]: DEBUG oslo.service.loopingcall [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.039578] env[62558]: DEBUG nova.compute.manager [-] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.039578] env[62558]: DEBUG nova.network.neutron [-] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.069474] env[62558]: DEBUG nova.network.neutron [-] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.190419] env[62558]: DEBUG nova.network.neutron [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.228922] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.229729] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.229931] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.230148] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.230294] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.230474] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.230649] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.230809] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.231081] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.231178] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.231361] env[62558]: DEBUG nova.virt.hardware [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.232543] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f085e4fc-0cb5-4382-86d1-2e408a2e0bab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.244546] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbc8e4c-aaa4-4af8-9cfa-d1f05bd89682 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.259043] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.269342] env[62558]: DEBUG oslo.service.loopingcall [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.269715] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 548.270287] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a08fb9a-3037-454c-b626-dc046cf86f96 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.288179] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.288415] env[62558]: ERROR nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Traceback (most recent call last): [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.driver.spawn(context, instance, image_meta, [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] vm_ref = self.build_virtual_machine(instance, [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.288415] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] for vif in network_info: [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self._sync_wrapper(fn, *args, **kwargs) [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.wait() [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self[:] = self._gt.wait() [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self._exit_event.wait() [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] result = hub.switch() [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.288885] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return self.greenlet.switch() [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] result = function(*args, **kwargs) [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] return func(*args, **kwargs) [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise e [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] nwinfo = self.network_api.allocate_for_instance( [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] created_port_ids = self._update_ports_for_instance( [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] with excutils.save_and_reraise_exception(): [ 548.291313] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] self.force_reraise() [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise self.value [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] updated_port = self._update_port( [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] _ensure_no_port_binding_failure(port) [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] raise exception.PortBindingFailed(port_id=port['id']) [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] nova.exception.PortBindingFailed: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. [ 548.291847] env[62558]: ERROR nova.compute.manager [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] [ 548.292927] env[62558]: DEBUG nova.compute.utils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.292927] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.221s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.293887] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Build of instance 99296a9b-151c-4b32-8e6a-ad0e28945aac was re-scheduled: Binding failed for port a8a4a427-c551-4893-abfe-62502b097be8, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.294423] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.294692] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.294866] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquired lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.295063] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.299892] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.299892] env[62558]: value = "task-1266494" [ 548.299892] env[62558]: _type = "Task" [ 548.299892] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.308367] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266494, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.432054] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5a7acb62-3d23-49f8-b76d-3beebc2b108d tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "d6eaf42d-dd83-4158-a507-9bc71cd1aa43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.770s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.572451] env[62558]: DEBUG nova.network.neutron [-] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.693074] env[62558]: INFO nova.compute.manager [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] [instance: 158e242b-3f2d-4cac-b3f3-c51da7f78bb1] Took 1.05 seconds to deallocate network for instance. [ 548.718062] env[62558]: DEBUG nova.network.neutron [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.818484] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266494, 'name': CreateVM_Task, 'duration_secs': 0.345472} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.818647] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 548.819122] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.819250] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.819736] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 548.819855] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722128fb-ec16-4d00-80b4-5e0baa2b890f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.825529] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 548.825529] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522204b1-33d1-3354-c1fc-1ef433dea9ca" [ 548.825529] env[62558]: _type = "Task" [ 548.825529] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.835207] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522204b1-33d1-3354-c1fc-1ef433dea9ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.862555] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.935545] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.074397] env[62558]: INFO nova.compute.manager [-] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Took 1.03 seconds to deallocate network for instance. [ 549.076514] env[62558]: DEBUG nova.compute.claims [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.076688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.176390] env[62558]: DEBUG nova.network.neutron [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.338205] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.347522] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522204b1-33d1-3354-c1fc-1ef433dea9ca, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.349031] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 06df57a7-2c53-4f79-bec4-e46cfe1dca63 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 549.349235] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d02e6383-42d9-40db-89a3-ba8b5595b95b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 549.350914] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.351224] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 549.351466] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.351609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.351779] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 549.352259] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba220329-bdcb-4d7a-8f93-27ff0a22e20a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.365577] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 549.366169] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 549.366977] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32b31a3a-a8b3-434e-aca5-56383d7de082 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.372998] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 549.372998] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523ecd13-5aa2-2906-e038-552a4a42cab3" [ 549.372998] env[62558]: _type = "Task" [ 549.372998] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.387319] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523ecd13-5aa2-2906-e038-552a4a42cab3, 'name': SearchDatastore_Task, 'duration_secs': 0.008292} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.387988] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e1db2c6-49df-4592-a4e5-898d34f213ec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.394850] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 549.394850] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f89bc7-3d82-bee9-88ed-e1a0b51a086f" [ 549.394850] env[62558]: _type = "Task" [ 549.394850] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.405233] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f89bc7-3d82-bee9-88ed-e1a0b51a086f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.458640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.680645] env[62558]: DEBUG oslo_concurrency.lockutils [req-bf697b73-6298-4ea9-9516-7b86d339b036 req-4865ebef-4ea4-49d5-b218-32517924d009 service nova] Releasing lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.736898] env[62558]: INFO nova.scheduler.client.report [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Deleted allocations for instance 158e242b-3f2d-4cac-b3f3-c51da7f78bb1 [ 549.849201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Releasing lock "refresh_cache-99296a9b-151c-4b32-8e6a-ad0e28945aac" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.849473] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.849671] env[62558]: DEBUG nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.849837] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.853432] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 158e242b-3f2d-4cac-b3f3-c51da7f78bb1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 549.882822] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.911768] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f89bc7-3d82-bee9-88ed-e1a0b51a086f, 'name': SearchDatastore_Task, 'duration_secs': 0.022839} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.912137] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.913679] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 549.913679] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7355067c-e736-45e0-95b3-f14426a721a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.922539] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 549.922539] env[62558]: value = "task-1266495" [ 549.922539] env[62558]: _type = "Task" [ 549.922539] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.934463] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.135919] env[62558]: DEBUG nova.compute.manager [req-16c84518-8d0a-4880-aedc-363ce0264a46 req-31fb2bdf-cdbd-4735-b7eb-e1cb8b1a9740 service nova] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Received event network-vif-deleted-6996c67c-5f06-4713-aced-7d64fd17e0f1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.248132] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0326e14-cbc0-49d0-a1d5-d30055d14614 tempest-ServerDiagnosticsTest-2062788967 tempest-ServerDiagnosticsTest-2062788967-project-member] Lock "158e242b-3f2d-4cac-b3f3-c51da7f78bb1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.595s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.357935] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 99296a9b-151c-4b32-8e6a-ad0e28945aac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.358967] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9f923c30-6c68-4e5d-be8a-4f18e996a999 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.358967] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 7d66f026-8b3e-4670-ada6-0109d99bb984 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.358967] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance b00d6153-c94f-4605-9b97-9a3a813ee93b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.358967] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 3c173701-9f61-4561-ac3c-7ac3de4101d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.359127] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance b08cb5a2-56d1-43da-ae4a-684077dd517f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.384304] env[62558]: DEBUG nova.network.neutron [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.435635] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439469} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.435917] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 550.436637] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 550.436637] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6397fdb-86b4-4b94-a778-169b55b4e5f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.451302] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 550.451302] env[62558]: value = "task-1266496" [ 550.451302] env[62558]: _type = "Task" [ 550.451302] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.465833] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.659529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquiring lock "5a1f074f-8e43-42f9-864d-976cdaa4639c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.659529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "5a1f074f-8e43-42f9-864d-976cdaa4639c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.752944] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.864038] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9316facd-0c95-4720-b0d3-c0a39df81619 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.889024] env[62558]: INFO nova.compute.manager [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 99296a9b-151c-4b32-8e6a-ad0e28945aac] Took 1.04 seconds to deallocate network for instance. [ 550.963124] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064318} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.963124] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.963661] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbf30c7-6082-47d8-874e-07d751c578a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.992746] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.994026] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7015c3a9-318e-4594-88e4-505e0834c175 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.014710] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 551.014710] env[62558]: value = "task-1266497" [ 551.014710] env[62558]: _type = "Task" [ 551.014710] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.026207] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266497, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.292697] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.368460] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f52921c6-abbc-4a27-a451-e1cb4f5dfc48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.526686] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266497, 'name': ReconfigVM_Task, 'duration_secs': 0.306218} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.528926] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b/d02e6383-42d9-40db-89a3-ba8b5595b95b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.528926] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-872a27f5-6a3d-430d-9298-896b84995078 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.534148] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 551.534148] env[62558]: value = "task-1266498" [ 551.534148] env[62558]: _type = "Task" [ 551.534148] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.543044] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266498, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.870801] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5fc9a508-323d-4f1c-ae46-d15f340f6b73 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.935241] env[62558]: INFO nova.scheduler.client.report [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Deleted allocations for instance 99296a9b-151c-4b32-8e6a-ad0e28945aac [ 552.049354] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266498, 'name': Rename_Task, 'duration_secs': 0.130747} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.049827] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 552.050220] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f459930-ed8c-4a0e-b256-53066c0266a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.057084] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Waiting for the task: (returnval){ [ 552.057084] env[62558]: value = "task-1266499" [ 552.057084] env[62558]: _type = "Task" [ 552.057084] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.067542] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.375310] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5845d273-b18a-447a-ad2d-a011bfcc2423 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.445766] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7686bada-e150-4867-b03a-658edca360c5 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "99296a9b-151c-4b32-8e6a-ad0e28945aac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.522s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.570289] env[62558]: DEBUG oslo_vmware.api [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Task: {'id': task-1266499, 'name': PowerOnVM_Task, 'duration_secs': 0.498384} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.570772] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 552.572393] env[62558]: DEBUG nova.compute.manager [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.572393] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ea959a-986e-4742-8826-228c3b07d112 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.830710] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquiring lock "1513843d-0a96-4b56-86bb-3c39ab4b3692" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.830710] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "1513843d-0a96-4b56-86bb-3c39ab4b3692" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.880136] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.950079] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.090700] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.387040] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f4f7e2c9-229c-4547-b674-5a5219a28151 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.487134] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.536756] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.537014] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.577435] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquiring lock "2e1410c0-8aff-429c-8bbb-46d9bc01de32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.577681] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "2e1410c0-8aff-429c-8bbb-46d9bc01de32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.891509] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 813e287c-9d7f-4e9c-9216-986f631b6eb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.358149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquiring lock "95348a45-f517-4e21-8969-57c978985984" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.358149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "95348a45-f517-4e21-8969-57c978985984" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.394898] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.395189] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 554.395307] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 554.837122] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf04d8f-46c6-4bb0-9fd9-e2b4f1e803ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.849744] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71190cd4-2168-4b9d-bb6f-9e0b11a14f18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.891684] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f8d86b-2e32-45fc-a5a2-fd18cfa63618 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.898935] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8022024-8afa-48dc-a85a-80c89be759e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.914902] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.082690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.082690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.082690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "d02e6383-42d9-40db-89a3-ba8b5595b95b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.082690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.082852] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.085829] env[62558]: INFO nova.compute.manager [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Terminating instance [ 555.094462] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "refresh_cache-d02e6383-42d9-40db-89a3-ba8b5595b95b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.094462] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquired lock "refresh_cache-d02e6383-42d9-40db-89a3-ba8b5595b95b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.094462] env[62558]: DEBUG nova.network.neutron [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.419144] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.716019] env[62558]: DEBUG nova.network.neutron [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.928652] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 555.929018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.638s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.929393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.854s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.982870] env[62558]: DEBUG nova.network.neutron [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.492842] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Releasing lock "refresh_cache-d02e6383-42d9-40db-89a3-ba8b5595b95b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.493431] env[62558]: DEBUG nova.compute.manager [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.493932] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.499171] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e156c8-c994-4f92-b806-7c292ba891b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.508620] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 556.508822] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53c1bb2c-7efe-4792-9d0b-f1e2b5b3e232 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.519327] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 556.519327] env[62558]: value = "task-1266500" [ 556.519327] env[62558]: _type = "Task" [ 556.519327] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.532743] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.869305] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ed6c51-8d1b-4953-927e-fcb8da602910 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.882656] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0325f2-8d2c-4e24-9712-cd59895d992c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.921533] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc53c43-2bee-4452-bf1f-8e7936ff7c48 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.930722] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4e4e55-afec-429e-8962-4dfa901be4b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.945378] env[62558]: DEBUG nova.compute.provider_tree [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.984945] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "290e15dc-4ee0-4595-8412-d610895aeb46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.984945] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "290e15dc-4ee0-4595-8412-d610895aeb46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.031426] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266500, 'name': PowerOffVM_Task, 'duration_secs': 0.137528} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.031704] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 557.031868] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 557.035084] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea91651e-4549-4f37-898b-409fca41183b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.061517] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 557.061855] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 557.061855] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleting the datastore file [datastore2] d02e6383-42d9-40db-89a3-ba8b5595b95b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 557.062106] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-092c0ac7-f118-4e07-81fa-caf70194f19c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.069361] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for the task: (returnval){ [ 557.069361] env[62558]: value = "task-1266502" [ 557.069361] env[62558]: _type = "Task" [ 557.069361] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.078119] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266502, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.449101] env[62558]: DEBUG nova.scheduler.client.report [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.586092] env[62558]: DEBUG oslo_vmware.api [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Task: {'id': task-1266502, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125308} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.586399] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.589622] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 557.589622] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.589622] env[62558]: INFO nova.compute.manager [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 557.589622] env[62558]: DEBUG oslo.service.loopingcall [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.589622] env[62558]: DEBUG nova.compute.manager [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.590288] env[62558]: DEBUG nova.network.neutron [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.647650] env[62558]: DEBUG nova.network.neutron [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.957992] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.958968] env[62558]: ERROR nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Traceback (most recent call last): [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.driver.spawn(context, instance, image_meta, [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] vm_ref = self.build_virtual_machine(instance, [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.958968] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] for vif in network_info: [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self._sync_wrapper(fn, *args, **kwargs) [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.wait() [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self[:] = self._gt.wait() [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self._exit_event.wait() [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] result = hub.switch() [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.959489] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return self.greenlet.switch() [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] result = function(*args, **kwargs) [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] return func(*args, **kwargs) [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise e [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] nwinfo = self.network_api.allocate_for_instance( [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] created_port_ids = self._update_ports_for_instance( [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] with excutils.save_and_reraise_exception(): [ 557.963528] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] self.force_reraise() [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise self.value [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] updated_port = self._update_port( [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] _ensure_no_port_binding_failure(port) [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] raise exception.PortBindingFailed(port_id=port['id']) [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] nova.exception.PortBindingFailed: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. [ 557.969964] env[62558]: ERROR nova.compute.manager [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] [ 557.970632] env[62558]: DEBUG nova.compute.utils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.970632] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.407s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.970632] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Build of instance 9f923c30-6c68-4e5d-be8a-4f18e996a999 was re-scheduled: Binding failed for port ce0a2b84-a054-4344-9de1-274fd798155b, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.970632] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.970954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.970988] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquired lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.974557] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.151519] env[62558]: DEBUG nova.network.neutron [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.507976] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.657931] env[62558]: INFO nova.compute.manager [-] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Took 1.07 seconds to deallocate network for instance. [ 558.821133] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6dfa07-b4a6-49f9-9bb3-96e96efc8db2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.830090] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0366ab79-7460-4645-8a97-7c360f689d6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.863235] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.864663] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356de27f-01a2-4f28-af2a-60a2bd733cc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.872832] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5bdb50-28cf-4664-ad1f-00ac3557a359 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.889471] env[62558]: DEBUG nova.compute.provider_tree [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.165175] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.181799] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquiring lock "49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.182445] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.370713] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Releasing lock "refresh_cache-9f923c30-6c68-4e5d-be8a-4f18e996a999" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.370713] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.370713] env[62558]: DEBUG nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.370713] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.395311] env[62558]: DEBUG nova.scheduler.client.report [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.418893] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.901749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.904100] env[62558]: ERROR nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Traceback (most recent call last): [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.driver.spawn(context, instance, image_meta, [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] vm_ref = self.build_virtual_machine(instance, [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.904100] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] for vif in network_info: [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self._sync_wrapper(fn, *args, **kwargs) [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.wait() [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self[:] = self._gt.wait() [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self._exit_event.wait() [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] result = hub.switch() [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.904489] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return self.greenlet.switch() [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] result = function(*args, **kwargs) [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] return func(*args, **kwargs) [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise e [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] nwinfo = self.network_api.allocate_for_instance( [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] created_port_ids = self._update_ports_for_instance( [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] with excutils.save_and_reraise_exception(): [ 559.904862] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] self.force_reraise() [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise self.value [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] updated_port = self._update_port( [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] _ensure_no_port_binding_failure(port) [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] raise exception.PortBindingFailed(port_id=port['id']) [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] nova.exception.PortBindingFailed: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. [ 559.905267] env[62558]: ERROR nova.compute.manager [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] [ 559.905604] env[62558]: DEBUG nova.compute.utils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.912122] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.961s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.912122] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Build of instance 7d66f026-8b3e-4670-ada6-0109d99bb984 was re-scheduled: Binding failed for port 4901996c-557b-4324-90fc-d0762fb23d46, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.912122] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.912122] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquiring lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.912475] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Acquired lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.912475] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 559.926338] env[62558]: DEBUG nova.network.neutron [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.241151] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquiring lock "d46bdbc6-e691-44b7-8e9c-3bedb1275e93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.243064] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "d46bdbc6-e691-44b7-8e9c-3bedb1275e93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.429600] env[62558]: INFO nova.compute.manager [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: 9f923c30-6c68-4e5d-be8a-4f18e996a999] Took 1.06 seconds to deallocate network for instance. [ 560.501351] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.928765] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.942355] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b394f0-42a8-46b5-8217-6d161d9f0091 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.957297] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b883622f-c402-4ad5-aeaf-8003ed763624 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.012162] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da73c928-ebe0-4ff5-98b2-133d749eb307 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.022357] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667ac10d-e055-405e-9a97-9aeefdaae0bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.042463] env[62558]: DEBUG nova.compute.provider_tree [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.432128] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Releasing lock "refresh_cache-7d66f026-8b3e-4670-ada6-0109d99bb984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.432203] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 561.432412] env[62558]: DEBUG nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.432585] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.479525] env[62558]: INFO nova.scheduler.client.report [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Deleted allocations for instance 9f923c30-6c68-4e5d-be8a-4f18e996a999 [ 561.487556] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.545390] env[62558]: DEBUG nova.scheduler.client.report [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.990233] env[62558]: DEBUG nova.network.neutron [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.991635] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2b5042bd-982b-4a17-afa2-c80515085efe tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "9f923c30-6c68-4e5d-be8a-4f18e996a999" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.723s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.050659] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.143s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.051331] env[62558]: ERROR nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Traceback (most recent call last): [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.driver.spawn(context, instance, image_meta, [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] vm_ref = self.build_virtual_machine(instance, [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.051331] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] for vif in network_info: [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self._sync_wrapper(fn, *args, **kwargs) [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.wait() [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self[:] = self._gt.wait() [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self._exit_event.wait() [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] result = hub.switch() [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.053278] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return self.greenlet.switch() [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] result = function(*args, **kwargs) [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] return func(*args, **kwargs) [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise e [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] nwinfo = self.network_api.allocate_for_instance( [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] created_port_ids = self._update_ports_for_instance( [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] with excutils.save_and_reraise_exception(): [ 562.053653] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] self.force_reraise() [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise self.value [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] updated_port = self._update_port( [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] _ensure_no_port_binding_failure(port) [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] raise exception.PortBindingFailed(port_id=port['id']) [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] nova.exception.PortBindingFailed: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. [ 562.054101] env[62558]: ERROR nova.compute.manager [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] [ 562.054432] env[62558]: DEBUG nova.compute.utils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.056722] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.805s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.056722] env[62558]: INFO nova.compute.claims [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.061681] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Build of instance b00d6153-c94f-4605-9b97-9a3a813ee93b was re-scheduled: Binding failed for port 2864207b-caee-484d-808e-9ddbe63de986, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 562.062169] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 562.062387] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquiring lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.062524] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Acquired lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.062673] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.500603] env[62558]: INFO nova.compute.manager [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] [instance: 7d66f026-8b3e-4670-ada6-0109d99bb984] Took 1.07 seconds to deallocate network for instance. [ 562.503962] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.721199] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.887681] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.033199] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.393932] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Releasing lock "refresh_cache-b00d6153-c94f-4605-9b97-9a3a813ee93b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.393932] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.393932] env[62558]: DEBUG nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.394081] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.457826] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.549966] env[62558]: INFO nova.scheduler.client.report [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Deleted allocations for instance 7d66f026-8b3e-4670-ada6-0109d99bb984 [ 563.577627] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80ef35f-a25f-421d-be5c-bef3c488728a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.586531] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b040a6d8-aaa0-4207-b196-6369a444165f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.627833] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97733195-bb87-42d1-a6c8-881ed821d2d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.636185] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f48f3a1-e555-49ed-bc1b-9ffc6ef07b02 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.652875] env[62558]: DEBUG nova.compute.provider_tree [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.963362] env[62558]: DEBUG nova.network.neutron [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.065199] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8844b877-3b73-4b80-bc33-f7ae4c60ff40 tempest-ServersAdminTestJSON-1077592274 tempest-ServersAdminTestJSON-1077592274-project-member] Lock "7d66f026-8b3e-4670-ada6-0109d99bb984" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.120s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.156570] env[62558]: DEBUG nova.scheduler.client.report [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.319880] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquiring lock "8e7dcc93-2079-4857-8f18-513958ef4e0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.320335] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "8e7dcc93-2079-4857-8f18-513958ef4e0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.467862] env[62558]: INFO nova.compute.manager [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] [instance: b00d6153-c94f-4605-9b97-9a3a813ee93b] Took 1.07 seconds to deallocate network for instance. [ 564.573639] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.663267] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.663267] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.667974] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.052s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.670161] env[62558]: INFO nova.compute.claims [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.108667] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.151212] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquiring lock "9274ce52-49af-48df-bcd1-269b8241d93f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.151501] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "9274ce52-49af-48df-bcd1-269b8241d93f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.177186] env[62558]: DEBUG nova.compute.utils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.183140] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 565.312354] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "88cee397-bcbd-428c-b203-e5d9b6acbbb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.312625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "88cee397-bcbd-428c-b203-e5d9b6acbbb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.344896] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ef19262f-77be-470a-b510-29b8016da5a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.345195] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ef19262f-77be-470a-b510-29b8016da5a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.508041] env[62558]: INFO nova.scheduler.client.report [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Deleted allocations for instance b00d6153-c94f-4605-9b97-9a3a813ee93b [ 565.685435] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.019812] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4d3a1091-20a8-415b-83e5-52590d1d7be7 tempest-ServerDiagnosticsNegativeTest-2059032827 tempest-ServerDiagnosticsNegativeTest-2059032827-project-member] Lock "b00d6153-c94f-4605-9b97-9a3a813ee93b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.989s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.120450] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e2e64a-0554-4074-9e3c-fb915cbc6568 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.132320] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43abf4fc-8a2c-4330-af3c-9a1538b9b71d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.168720] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721287de-68b8-4bbf-a6d7-711eef0654ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.177177] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c700bb8c-cd1a-4d4b-9dd6-6812faf009c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.191512] env[62558]: DEBUG nova.compute.provider_tree [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.522905] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.705339] env[62558]: DEBUG nova.scheduler.client.report [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.708180] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.745858] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.745998] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.746162] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.746357] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.746499] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.746640] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.746839] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.747253] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.747253] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.747371] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.747531] env[62558]: DEBUG nova.virt.hardware [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.748787] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad16d09-d4bf-4308-86df-b33ce21c9a0c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.759707] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba23344-f380-4072-bc48-71ef7f1cc23e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.776462] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 566.784207] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Creating folder: Project (eb5281fa984f45a2ab4c3136c640e897). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 566.784207] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c13ecb0-d0a8-4066-b1f7-7de8431e7c64 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.793158] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Created folder: Project (eb5281fa984f45a2ab4c3136c640e897) in parent group-v272451. [ 566.793333] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Creating folder: Instances. Parent ref: group-v272460. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 566.793748] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f71d7db-424d-4b80-8361-fe70dabc0e67 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.802705] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Created folder: Instances in parent group-v272460. [ 566.803068] env[62558]: DEBUG oslo.service.loopingcall [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.803370] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 566.803658] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c00d2aaf-a2d5-44d5-8ed1-f91148bd8881 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.821240] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 566.821240] env[62558]: value = "task-1266505" [ 566.821240] env[62558]: _type = "Task" [ 566.821240] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.830160] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266505, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.045438] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.212223] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.212749] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.221181] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.067s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.221181] env[62558]: DEBUG nova.objects.instance [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 567.334217] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266505, 'name': CreateVM_Task, 'duration_secs': 0.299288} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.334365] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 567.335582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.335582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.335582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 567.335582] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-895261ef-4d33-4319-bd37-eeafbe1b6e1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.340111] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 567.340111] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb16c-cf95-39f2-165b-179927fbbfb5" [ 567.340111] env[62558]: _type = "Task" [ 567.340111] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.348043] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb16c-cf95-39f2-165b-179927fbbfb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.718341] env[62558]: DEBUG nova.compute.utils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.719790] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.719965] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 567.851741] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb16c-cf95-39f2-165b-179927fbbfb5, 'name': SearchDatastore_Task, 'duration_secs': 0.008319} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.852066] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.852302] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 567.852531] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.852673] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.852844] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 567.853117] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8418653-9001-479a-bba7-a52cd5615b04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.862600] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 567.862600] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 567.863475] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-661d47ad-d00e-4b52-b305-a9cb64d8accf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.872127] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 567.872127] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523d44c4-fecc-f1e5-8538-5fa20917a262" [ 567.872127] env[62558]: _type = "Task" [ 567.872127] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.879417] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523d44c4-fecc-f1e5-8538-5fa20917a262, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.883081] env[62558]: DEBUG nova.policy [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8e4c35b45c64777bc690550450533f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a0dce7fbc2b4289bb428c99fea936e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.224377] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.232012] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c781e22-3203-4384-8388-f3ca5c3a28e8 tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.232976] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.448s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.233185] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.239609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.601s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.257848] env[62558]: INFO nova.scheduler.client.report [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Deleted allocations for instance 06df57a7-2c53-4f79-bec4-e46cfe1dca63 [ 568.383214] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523d44c4-fecc-f1e5-8538-5fa20917a262, 'name': SearchDatastore_Task, 'duration_secs': 0.018051} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.384105] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a65abd5f-2130-43de-b092-22dd8f085ad2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.391611] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 568.391611] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b7f06b-54ba-2819-6d64-83fc57f693ac" [ 568.391611] env[62558]: _type = "Task" [ 568.391611] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.403459] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b7f06b-54ba-2819-6d64-83fc57f693ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.643015] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquiring lock "2236b710-2141-49a2-b3f5-9ed71090e1db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.643015] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "2236b710-2141-49a2-b3f5-9ed71090e1db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.770732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4cd1281-9622-43d9-ba5b-46cd7c0c85e5 tempest-ServerDiagnosticsV248Test-757203038 tempest-ServerDiagnosticsV248Test-757203038-project-member] Lock "06df57a7-2c53-4f79-bec4-e46cfe1dca63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.959s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.910337] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b7f06b-54ba-2819-6d64-83fc57f693ac, 'name': SearchDatastore_Task, 'duration_secs': 0.015106} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.910610] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.910883] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 568.911148] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fceb3d8a-840f-4da4-85cc-e8c41aad4743 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.919933] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 568.919933] env[62558]: value = "task-1266506" [ 568.919933] env[62558]: _type = "Task" [ 568.919933] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.934733] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.161909] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Successfully created port: 8b26b7d8-1344-489e-90c0-b07a7bc7ba89 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.234266] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.280983] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.281250] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.281400] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.281576] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.281714] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.281866] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.282142] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.282313] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.282488] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.282668] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.282799] env[62558]: DEBUG nova.virt.hardware [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.283816] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a15a410-3d3f-406e-a898-73fb40a2555c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.288223] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f98faee-ca6f-4ca4-b869-5e1337299373 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.298177] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200bbf71-a17b-4bd5-8f7e-ffdf98d4fbbd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.302793] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef71db8-e830-4e7b-8397-32eef31c73c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.347490] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec586bb-8152-4751-b23e-fb5474dfa272 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.356821] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8cca87-d657-4df7-894c-05db2a8c4f70 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.373796] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "dca0a819-d11d-437e-9a4c-da4279a9d553" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.374115] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "dca0a819-d11d-437e-9a4c-da4279a9d553" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.375084] env[62558]: DEBUG nova.compute.provider_tree [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.430451] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484257} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.430708] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 569.430914] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 569.431185] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7ac68d3-7dbc-49ff-9dca-61b2e1cf36c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.437939] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 569.437939] env[62558]: value = "task-1266507" [ 569.437939] env[62558]: _type = "Task" [ 569.437939] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.450273] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.877844] env[62558]: DEBUG nova.scheduler.client.report [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.955289] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065907} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.955574] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 569.960022] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4032aa-3153-43f9-8f06-2658f287dd06 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.983747] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 569.984546] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae6f2d8f-8db7-434e-89b2-5e044845a8ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.005810] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 570.005810] env[62558]: value = "task-1266508" [ 570.005810] env[62558]: _type = "Task" [ 570.005810] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.014488] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.388024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.146s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.388024] env[62558]: ERROR nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Traceback (most recent call last): [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.driver.spawn(context, instance, image_meta, [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.388024] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] vm_ref = self.build_virtual_machine(instance, [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] for vif in network_info: [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self._sync_wrapper(fn, *args, **kwargs) [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.wait() [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self[:] = self._gt.wait() [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self._exit_event.wait() [ 570.388706] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] result = hub.switch() [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return self.greenlet.switch() [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] result = function(*args, **kwargs) [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] return func(*args, **kwargs) [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise e [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] nwinfo = self.network_api.allocate_for_instance( [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.389100] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] created_port_ids = self._update_ports_for_instance( [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] with excutils.save_and_reraise_exception(): [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] self.force_reraise() [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise self.value [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] updated_port = self._update_port( [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] _ensure_no_port_binding_failure(port) [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.389474] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] raise exception.PortBindingFailed(port_id=port['id']) [ 570.390738] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] nova.exception.PortBindingFailed: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. [ 570.390738] env[62558]: ERROR nova.compute.manager [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] [ 570.390738] env[62558]: DEBUG nova.compute.utils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.393191] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.316s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.397414] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Build of instance 3c173701-9f61-4561-ac3c-7ac3de4101d9 was re-scheduled: Binding failed for port 54187e76-94e3-4303-be6d-5241a92560bb, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 570.397414] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 570.397414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquiring lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.397414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Acquired lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.397703] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.521483] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266508, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.963589] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.030220] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266508, 'name': ReconfigVM_Task, 'duration_secs': 0.762887} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.033062] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 571.033062] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-834eb045-ada5-43f5-b2f1-ccb9e4bc068b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.041733] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 571.041733] env[62558]: value = "task-1266509" [ 571.041733] env[62558]: _type = "Task" [ 571.041733] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.059551] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266509, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.351245] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.400653] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b12c79-d57f-4e82-8f9a-a5bbf1d13841 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.414243] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b068e9ca-16a6-40ce-a18d-9d7a43370d47 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.453254] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aaaf076-3fa9-41ab-9e62-034ad197c708 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.463261] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5f2340-df54-463b-b401-d424d282e12e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.476454] env[62558]: DEBUG nova.compute.provider_tree [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.556975] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266509, 'name': Rename_Task, 'duration_secs': 0.139567} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.556975] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 571.556975] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb8cd469-37fc-40f6-a80f-b7937d2e0946 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.563634] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 571.563634] env[62558]: value = "task-1266510" [ 571.563634] env[62558]: _type = "Task" [ 571.563634] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.571852] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.853816] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Releasing lock "refresh_cache-3c173701-9f61-4561-ac3c-7ac3de4101d9" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.854092] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 571.854283] env[62558]: DEBUG nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.854450] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.900958] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.981223] env[62558]: DEBUG nova.scheduler.client.report [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.075367] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.406420] env[62558]: DEBUG nova.network.neutron [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.485182] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.485862] env[62558]: ERROR nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Traceback (most recent call last): [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.driver.spawn(context, instance, image_meta, [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] vm_ref = self.build_virtual_machine(instance, [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.485862] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] for vif in network_info: [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self._sync_wrapper(fn, *args, **kwargs) [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.wait() [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self[:] = self._gt.wait() [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self._exit_event.wait() [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] result = hub.switch() [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.486706] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return self.greenlet.switch() [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] result = function(*args, **kwargs) [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] return func(*args, **kwargs) [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise e [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] nwinfo = self.network_api.allocate_for_instance( [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] created_port_ids = self._update_ports_for_instance( [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] with excutils.save_and_reraise_exception(): [ 572.487590] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] self.force_reraise() [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise self.value [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] updated_port = self._update_port( [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] _ensure_no_port_binding_failure(port) [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] raise exception.PortBindingFailed(port_id=port['id']) [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] nova.exception.PortBindingFailed: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. [ 572.488324] env[62558]: ERROR nova.compute.manager [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] [ 572.489673] env[62558]: DEBUG nova.compute.utils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.491099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.032s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.493081] env[62558]: INFO nova.compute.claims [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.496188] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Build of instance b08cb5a2-56d1-43da-ae4a-684077dd517f was re-scheduled: Binding failed for port 6996c67c-5f06-4713-aced-7d64fd17e0f1, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.496732] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.497588] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.498738] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.498738] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.576208] env[62558]: DEBUG oslo_vmware.api [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266510, 'name': PowerOnVM_Task, 'duration_secs': 0.605745} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.576757] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 572.578098] env[62558]: INFO nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Took 5.87 seconds to spawn the instance on the hypervisor. [ 572.578098] env[62558]: DEBUG nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 572.578244] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e3cf7c-72cd-40fa-89f2-ceb1fab28d89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.916380] env[62558]: INFO nova.compute.manager [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] [instance: 3c173701-9f61-4561-ac3c-7ac3de4101d9] Took 1.06 seconds to deallocate network for instance. [ 573.056574] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.102720] env[62558]: INFO nova.compute.manager [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Took 31.88 seconds to build instance. [ 573.339582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "65e8971a-a149-4a11-92ad-7c5c8874530c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.339809] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "65e8971a-a149-4a11-92ad-7c5c8874530c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.384033] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.560960] env[62558]: ERROR nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 573.560960] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.560960] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.560960] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.560960] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.560960] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.560960] env[62558]: ERROR nova.compute.manager raise self.value [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.560960] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.560960] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.560960] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.561739] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.561739] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.561739] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 573.561739] env[62558]: ERROR nova.compute.manager [ 573.561739] env[62558]: Traceback (most recent call last): [ 573.561739] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.561739] env[62558]: listener.cb(fileno) [ 573.561739] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.561739] env[62558]: result = function(*args, **kwargs) [ 573.561739] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.561739] env[62558]: return func(*args, **kwargs) [ 573.561739] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.561739] env[62558]: raise e [ 573.561739] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.561739] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 573.561739] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.561739] env[62558]: created_port_ids = self._update_ports_for_instance( [ 573.561739] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.561739] env[62558]: with excutils.save_and_reraise_exception(): [ 573.561739] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.561739] env[62558]: self.force_reraise() [ 573.561739] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.561739] env[62558]: raise self.value [ 573.561739] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.561739] env[62558]: updated_port = self._update_port( [ 573.561739] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.561739] env[62558]: _ensure_no_port_binding_failure(port) [ 573.561739] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.561739] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.563716] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 573.563716] env[62558]: Removing descriptor: 22 [ 573.563716] env[62558]: ERROR nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Traceback (most recent call last): [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] yield resources [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.driver.spawn(context, instance, image_meta, [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.563716] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] vm_ref = self.build_virtual_machine(instance, [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] for vif in network_info: [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self._sync_wrapper(fn, *args, **kwargs) [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.wait() [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self[:] = self._gt.wait() [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self._exit_event.wait() [ 573.564350] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] result = hub.switch() [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self.greenlet.switch() [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] result = function(*args, **kwargs) [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return func(*args, **kwargs) [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise e [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] nwinfo = self.network_api.allocate_for_instance( [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.565381] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] created_port_ids = self._update_ports_for_instance( [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] with excutils.save_and_reraise_exception(): [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.force_reraise() [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise self.value [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] updated_port = self._update_port( [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] _ensure_no_port_binding_failure(port) [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.565802] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise exception.PortBindingFailed(port_id=port['id']) [ 573.566210] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 573.566210] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] [ 573.566210] env[62558]: INFO nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Terminating instance [ 573.572194] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquiring lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.575425] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquired lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.575425] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.575425] env[62558]: DEBUG nova.compute.manager [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Received event network-changed-8b26b7d8-1344-489e-90c0-b07a7bc7ba89 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.575425] env[62558]: DEBUG nova.compute.manager [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Refreshing instance network info cache due to event network-changed-8b26b7d8-1344-489e-90c0-b07a7bc7ba89. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 573.575425] env[62558]: DEBUG oslo_concurrency.lockutils [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] Acquiring lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.606174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-57b02ef4-c4a4-48df-bd14-e77f33a25b2a tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.169s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.886977] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-b08cb5a2-56d1-43da-ae4a-684077dd517f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.888130] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.888130] env[62558]: DEBUG nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.888130] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.940388] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.966032] env[62558]: INFO nova.scheduler.client.report [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Deleted allocations for instance 3c173701-9f61-4561-ac3c-7ac3de4101d9 [ 574.026545] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d57129-2f35-4d2b-850d-ee2588fd3560 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.037973] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59dbe43-c307-4109-9ec3-3d6b2108ab26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.074995] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15cf9c2-6455-4190-a75f-416ff9e7c829 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.085214] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0546d994-2d7c-4ebe-ab2a-0a3d82c5f9dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.102633] env[62558]: DEBUG nova.compute.provider_tree [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.110027] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.112525] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.355902] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.443074] env[62558]: DEBUG nova.network.neutron [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.480753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b2011e0d-fe1a-4316-b04f-c5592d516a30 tempest-VolumesAssistedSnapshotsTest-1179061681 tempest-VolumesAssistedSnapshotsTest-1179061681-project-member] Lock "3c173701-9f61-4561-ac3c-7ac3de4101d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.480s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.534314] env[62558]: INFO nova.compute.manager [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Rebuilding instance [ 574.606792] env[62558]: DEBUG nova.scheduler.client.report [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.615640] env[62558]: DEBUG nova.compute.manager [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 574.620132] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62ea74a-887b-4a38-a0c6-aac38324be07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.651816] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.860709] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Releasing lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.861185] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.861376] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.861715] env[62558]: DEBUG oslo_concurrency.lockutils [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] Acquired lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.861874] env[62558]: DEBUG nova.network.neutron [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Refreshing network info cache for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 574.863523] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6296d23d-e23f-465e-ab55-7d304bb36fa4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.878806] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332a6506-ff2b-48f0-a84d-531972549608 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.905100] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f52921c6-abbc-4a27-a451-e1cb4f5dfc48 could not be found. [ 574.905342] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.905626] env[62558]: INFO nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Took 0.04 seconds to destroy the instance on the hypervisor. [ 574.905768] env[62558]: DEBUG oslo.service.loopingcall [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.905994] env[62558]: DEBUG nova.compute.manager [-] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.906183] env[62558]: DEBUG nova.network.neutron [-] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.946531] env[62558]: DEBUG nova.network.neutron [-] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.950940] env[62558]: INFO nova.compute.manager [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: b08cb5a2-56d1-43da-ae4a-684077dd517f] Took 1.06 seconds to deallocate network for instance. [ 574.983058] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.130930] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.640s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.133071] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.147387] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.854s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.149967] env[62558]: INFO nova.compute.claims [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.156765] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 575.157981] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "b482dcca-ae32-425d-a097-f9af69785835" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.157981] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "b482dcca-ae32-425d-a097-f9af69785835" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.159961] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-509964bb-3a8d-45ea-9579-cc304648bd79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.176157] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 575.176157] env[62558]: value = "task-1266515" [ 575.176157] env[62558]: _type = "Task" [ 575.176157] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.189405] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.422987] env[62558]: DEBUG nova.network.neutron [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.452128] env[62558]: DEBUG nova.network.neutron [-] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.522862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.648406] env[62558]: DEBUG nova.compute.utils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.648661] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.649143] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.689738] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266515, 'name': PowerOffVM_Task, 'duration_secs': 0.162429} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.690426] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 575.690426] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.690999] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b324956-f292-4b48-890a-2719c47be38b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.698098] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 575.698362] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca89ad9b-ffb4-4cb9-8fae-7c42d077aaac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.723033] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 575.723262] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 575.723445] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Deleting the datastore file [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.723711] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46b6c282-50fb-442e-bb6c-8d19a1f753c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.734734] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 575.734734] env[62558]: value = "task-1266518" [ 575.734734] env[62558]: _type = "Task" [ 575.734734] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.742379] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.814354] env[62558]: DEBUG nova.network.neutron [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.880072] env[62558]: DEBUG nova.policy [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e63faf20d13b4b6eb274deb5dc7924c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62b80560ffed4c88822ad4c949e457ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.959147] env[62558]: INFO nova.compute.manager [-] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Took 1.05 seconds to deallocate network for instance. [ 575.961208] env[62558]: DEBUG nova.compute.claims [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.961512] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.990242] env[62558]: INFO nova.scheduler.client.report [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance b08cb5a2-56d1-43da-ae4a-684077dd517f [ 576.152625] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.246864] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239018} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.247153] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 576.247640] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 576.247640] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.312796] env[62558]: DEBUG nova.compute.manager [req-d452732c-c581-40f7-9e2b-79a1f871a80c req-63f7c428-1e1c-4fb8-ab20-926930f6d4fb service nova] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Received event network-vif-deleted-8b26b7d8-1344-489e-90c0-b07a7bc7ba89 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.317024] env[62558]: DEBUG oslo_concurrency.lockutils [req-3d6bdf79-d398-4fa8-9e04-bd0317188143 req-33f7fe20-cb58-48a8-a3ca-9117d0127f23 service nova] Releasing lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.359674] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Successfully created port: 2a94ba19-7663-49a9-90cf-06b74bb6525b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.502709] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a17e8905-a9d7-4eea-b7a8-8d24159283e2 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "b08cb5a2-56d1-43da-ae4a-684077dd517f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.576s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.597680] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c232b1-a226-4a2c-bd16-890a2331a149 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.605585] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ec61f9-0481-46d4-a8f1-1fbc298480ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.644606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e94ffbc-9261-42e4-80d4-807050461901 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.652429] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a763d50e-a6da-4d97-8074-3ff1d3209482 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.673969] env[62558]: DEBUG nova.compute.provider_tree [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.007017] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.177831] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.184890] env[62558]: DEBUG nova.scheduler.client.report [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.202577] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.202934] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.202934] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.203500] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.203744] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.203907] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.204169] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.204351] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.204508] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.204668] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.204838] env[62558]: DEBUG nova.virt.hardware [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.206734] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b515e717-a1f1-4b7d-bab9-216418b903fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.217231] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb722f55-8689-4ea6-a8d8-1e34ec71fbc0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.293230] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.293532] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.293788] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.293957] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.294176] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.294365] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.294604] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.294797] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.295007] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.295218] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.295432] env[62558]: DEBUG nova.virt.hardware [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.296436] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c067da26-ce32-404f-9cb0-c8b97f2fb7a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.305690] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f357d59-29d0-4385-ae78-5b2983290db2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.320594] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 577.326554] env[62558]: DEBUG oslo.service.loopingcall [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.326861] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 577.327156] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ce98062-2bc7-4a6d-b008-abf580caacab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.345203] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 577.345203] env[62558]: value = "task-1266519" [ 577.345203] env[62558]: _type = "Task" [ 577.345203] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.354232] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266519, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.528246] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.605069] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquiring lock "f8db751a-c495-467c-ae32-addb4387346d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.605519] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "f8db751a-c495-467c-ae32-addb4387346d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.694715] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.694715] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.700098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.606s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.700098] env[62558]: DEBUG nova.objects.instance [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 577.856442] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266519, 'name': CreateVM_Task, 'duration_secs': 0.314} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.856609] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 577.857028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.857184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.857504] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 577.857935] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd64105a-f9ba-4e14-88af-309ebd0fd4ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.863919] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 577.863919] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f9c0e-be86-6224-a627-9b7556e3b35f" [ 577.863919] env[62558]: _type = "Task" [ 577.863919] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.871938] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f9c0e-be86-6224-a627-9b7556e3b35f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.202132] env[62558]: DEBUG nova.compute.utils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.203667] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.203835] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.359480] env[62558]: DEBUG nova.policy [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c32356f84bae49ef95f4395daee5e91e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fafd1092c04e4472849e56b06c9f6a10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.379441] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f9c0e-be86-6224-a627-9b7556e3b35f, 'name': SearchDatastore_Task, 'duration_secs': 0.009457} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.380219] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.380219] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 578.380397] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.381037] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.381037] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.382175] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49cddcb7-8fb0-4511-9bee-d4eb0633397b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.391921] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.392354] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 578.392841] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3286c8cd-8752-4650-be92-10b412ad63e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.398151] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 578.398151] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a5198f-8ab6-a43f-35c7-5f0a24a80922" [ 578.398151] env[62558]: _type = "Task" [ 578.398151] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.405998] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a5198f-8ab6-a43f-35c7-5f0a24a80922, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.713202] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.713202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-baa6bca4-ee6b-4cd3-9a78-90796a4bf3e5 tempest-ServersAdmin275Test-2115638117 tempest-ServersAdmin275Test-2115638117-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.714533] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.228s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.716250] env[62558]: INFO nova.compute.claims [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.908470] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a5198f-8ab6-a43f-35c7-5f0a24a80922, 'name': SearchDatastore_Task, 'duration_secs': 0.008418} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.909308] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-152d4dbe-0b75-49f1-b4c1-fe1c6e8a3366 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.915568] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 578.915568] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5246abb0-2f3d-865e-f915-201cabd41cc7" [ 578.915568] env[62558]: _type = "Task" [ 578.915568] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.923934] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5246abb0-2f3d-865e-f915-201cabd41cc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.336087] env[62558]: ERROR nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 579.336087] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.336087] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.336087] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.336087] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.336087] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.336087] env[62558]: ERROR nova.compute.manager raise self.value [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.336087] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.336087] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.336087] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.336636] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.336636] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.336636] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 579.336636] env[62558]: ERROR nova.compute.manager [ 579.336636] env[62558]: Traceback (most recent call last): [ 579.336636] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.336636] env[62558]: listener.cb(fileno) [ 579.336636] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.336636] env[62558]: result = function(*args, **kwargs) [ 579.336636] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.336636] env[62558]: return func(*args, **kwargs) [ 579.336636] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.336636] env[62558]: raise e [ 579.336636] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.336636] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 579.336636] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.336636] env[62558]: created_port_ids = self._update_ports_for_instance( [ 579.336636] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.336636] env[62558]: with excutils.save_and_reraise_exception(): [ 579.336636] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.336636] env[62558]: self.force_reraise() [ 579.336636] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.336636] env[62558]: raise self.value [ 579.336636] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.336636] env[62558]: updated_port = self._update_port( [ 579.336636] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.336636] env[62558]: _ensure_no_port_binding_failure(port) [ 579.336636] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.336636] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.337619] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 579.337619] env[62558]: Removing descriptor: 18 [ 579.337619] env[62558]: ERROR nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Traceback (most recent call last): [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] yield resources [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.driver.spawn(context, instance, image_meta, [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.337619] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] vm_ref = self.build_virtual_machine(instance, [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] for vif in network_info: [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self._sync_wrapper(fn, *args, **kwargs) [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.wait() [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self[:] = self._gt.wait() [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self._exit_event.wait() [ 579.338279] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] result = hub.switch() [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self.greenlet.switch() [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] result = function(*args, **kwargs) [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return func(*args, **kwargs) [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise e [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] nwinfo = self.network_api.allocate_for_instance( [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.338704] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] created_port_ids = self._update_ports_for_instance( [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] with excutils.save_and_reraise_exception(): [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.force_reraise() [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise self.value [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] updated_port = self._update_port( [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] _ensure_no_port_binding_failure(port) [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.339139] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise exception.PortBindingFailed(port_id=port['id']) [ 579.339493] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 579.339493] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] [ 579.339493] env[62558]: INFO nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Terminating instance [ 579.340259] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquiring lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.340645] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquired lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.340970] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.433014] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5246abb0-2f3d-865e-f915-201cabd41cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.008627} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.433014] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.433014] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 579.433014] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0516a994-e1d2-443f-9d92-d3b41f7156db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.440688] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 579.440688] env[62558]: value = "task-1266522" [ 579.440688] env[62558]: _type = "Task" [ 579.440688] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.448629] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.538913] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Successfully created port: c83e936a-60b1-40bd-ba75-3d7c01abc582 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.640996] env[62558]: DEBUG nova.compute.manager [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Received event network-changed-2a94ba19-7663-49a9-90cf-06b74bb6525b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.641293] env[62558]: DEBUG nova.compute.manager [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Refreshing instance network info cache due to event network-changed-2a94ba19-7663-49a9-90cf-06b74bb6525b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.641527] env[62558]: DEBUG oslo_concurrency.lockutils [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] Acquiring lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.727956] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.773780] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.776685] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.776685] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.776685] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.776685] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.776685] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.777027] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.777027] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.778502] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.778502] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.778502] env[62558]: DEBUG nova.virt.hardware [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.780388] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe209a2-b8f1-4a3a-8c0f-055298f572eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.798199] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029c4427-1d4a-4a2e-84af-f7fb47152d19 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.886112] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.954563] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266522, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.095042] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquiring lock "9b49513e-4c01-4664-be2c-6058a357b0db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.095290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "9b49513e-4c01-4664-be2c-6058a357b0db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.140254] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.213688] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020876df-e5f7-4b4d-a481-b0d856d03888 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.226776] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b95036-8a58-4a66-80b3-18cada92900a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.254101] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5973e5c-09eb-43ec-8bb1-0989b572b2ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.262111] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b57855c-c0af-4953-8e07-3717af588e6d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.274924] env[62558]: DEBUG nova.compute.provider_tree [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.459510] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529068} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.461323] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 580.461323] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 580.461323] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69ad4060-954c-44e5-bdaf-6b2a0c7a181c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.466605] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 580.466605] env[62558]: value = "task-1266523" [ 580.466605] env[62558]: _type = "Task" [ 580.466605] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.475681] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.646284] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Releasing lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.648099] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.648099] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.648099] env[62558]: DEBUG oslo_concurrency.lockutils [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] Acquired lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.648099] env[62558]: DEBUG nova.network.neutron [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Refreshing network info cache for port 2a94ba19-7663-49a9-90cf-06b74bb6525b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.654122] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e944bb6a-da80-422a-82eb-b3fe1adb72d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.664605] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d2a676-2737-482f-9ca4-9bc0e8108013 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.691764] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5fc9a508-323d-4f1c-ae46-d15f340f6b73 could not be found. [ 580.692150] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.692464] env[62558]: INFO nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Took 0.05 seconds to destroy the instance on the hypervisor. [ 580.692794] env[62558]: DEBUG oslo.service.loopingcall [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.693123] env[62558]: DEBUG nova.compute.manager [-] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.694177] env[62558]: DEBUG nova.network.neutron [-] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.734818] env[62558]: DEBUG nova.network.neutron [-] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.783621] env[62558]: DEBUG nova.scheduler.client.report [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.790790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.790790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.978369] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064081} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.978628] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 580.979616] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91743f27-b02d-4afd-a82f-469c21db19be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.997658] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "2888fde6-8a1b-41ce-a482-cb99acd74a6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.997835] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "2888fde6-8a1b-41ce-a482-cb99acd74a6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.006872] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 581.007473] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfd27107-a648-47b5-af86-deece8868347 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.028763] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 581.028763] env[62558]: value = "task-1266524" [ 581.028763] env[62558]: _type = "Task" [ 581.028763] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.038826] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266524, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.237389] env[62558]: DEBUG nova.network.neutron [-] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.291341] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.295141] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 581.298768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.134s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.300387] env[62558]: DEBUG nova.objects.instance [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lazy-loading 'resources' on Instance uuid d02e6383-42d9-40db-89a3-ba8b5595b95b {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 581.352159] env[62558]: DEBUG nova.network.neutron [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.542889] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266524, 'name': ReconfigVM_Task, 'duration_secs': 0.272232} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.545548] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619/9316facd-0c95-4720-b0d3-c0a39df81619.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 581.545548] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4268b798-ff7c-41b1-b0eb-3b1599109a26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.558042] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 581.558042] env[62558]: value = "task-1266525" [ 581.558042] env[62558]: _type = "Task" [ 581.558042] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.565624] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266525, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.669991] env[62558]: DEBUG nova.network.neutron [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.741520] env[62558]: INFO nova.compute.manager [-] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Took 1.05 seconds to deallocate network for instance. [ 581.747318] env[62558]: DEBUG nova.compute.claims [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.747318] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.805491] env[62558]: DEBUG nova.compute.utils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.811423] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.811423] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.873290] env[62558]: ERROR nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 581.873290] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.873290] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.873290] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.873290] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.873290] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.873290] env[62558]: ERROR nova.compute.manager raise self.value [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.873290] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.873290] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.873290] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.873976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.873976] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.873976] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 581.873976] env[62558]: ERROR nova.compute.manager [ 581.873976] env[62558]: Traceback (most recent call last): [ 581.873976] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.873976] env[62558]: listener.cb(fileno) [ 581.873976] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.873976] env[62558]: result = function(*args, **kwargs) [ 581.873976] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.873976] env[62558]: return func(*args, **kwargs) [ 581.873976] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.873976] env[62558]: raise e [ 581.873976] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.873976] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 581.873976] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.873976] env[62558]: created_port_ids = self._update_ports_for_instance( [ 581.873976] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.873976] env[62558]: with excutils.save_and_reraise_exception(): [ 581.873976] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.873976] env[62558]: self.force_reraise() [ 581.873976] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.873976] env[62558]: raise self.value [ 581.873976] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.873976] env[62558]: updated_port = self._update_port( [ 581.873976] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.873976] env[62558]: _ensure_no_port_binding_failure(port) [ 581.873976] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.873976] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.874880] env[62558]: nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 581.874880] env[62558]: Removing descriptor: 22 [ 581.874880] env[62558]: ERROR nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Traceback (most recent call last): [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] yield resources [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.driver.spawn(context, instance, image_meta, [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.874880] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] vm_ref = self.build_virtual_machine(instance, [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] for vif in network_info: [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self._sync_wrapper(fn, *args, **kwargs) [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.wait() [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self[:] = self._gt.wait() [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self._exit_event.wait() [ 581.875310] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] result = hub.switch() [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self.greenlet.switch() [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] result = function(*args, **kwargs) [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return func(*args, **kwargs) [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise e [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] nwinfo = self.network_api.allocate_for_instance( [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.875806] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] created_port_ids = self._update_ports_for_instance( [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] with excutils.save_and_reraise_exception(): [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.force_reraise() [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise self.value [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] updated_port = self._update_port( [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] _ensure_no_port_binding_failure(port) [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.876686] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise exception.PortBindingFailed(port_id=port['id']) [ 581.877084] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 581.877084] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] [ 581.877084] env[62558]: INFO nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Terminating instance [ 581.879859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquiring lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.879859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquired lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.879859] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.919024] env[62558]: DEBUG nova.policy [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b177e8a629d2494aaf0497b520408909', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50cccd70acb04a4aa4691260e3437a94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.068275] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266525, 'name': Rename_Task, 'duration_secs': 0.132325} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.068275] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 582.068275] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88125f8e-038e-40a1-8375-9aa94d73d0ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.074965] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 582.074965] env[62558]: value = "task-1266527" [ 582.074965] env[62558]: _type = "Task" [ 582.074965] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.088366] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.177538] env[62558]: DEBUG oslo_concurrency.lockutils [req-5f954fd2-6db0-4410-acc3-1442369a63b2 req-3f28053e-f75f-45df-896b-3f5331cf11d9 service nova] Releasing lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.201923] env[62558]: DEBUG nova.compute.manager [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Received event network-vif-deleted-2a94ba19-7663-49a9-90cf-06b74bb6525b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.202156] env[62558]: DEBUG nova.compute.manager [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Received event network-changed-c83e936a-60b1-40bd-ba75-3d7c01abc582 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.202365] env[62558]: DEBUG nova.compute.manager [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Refreshing instance network info cache due to event network-changed-c83e936a-60b1-40bd-ba75-3d7c01abc582. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 582.202590] env[62558]: DEBUG oslo_concurrency.lockutils [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] Acquiring lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.285071] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff91563-2911-4293-9a51-7fe4b910dbcd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.292960] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f03e14c-502a-497a-a30b-a996a969624b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.328019] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 582.329486] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e2102b-4e95-4162-943c-8352bd120709 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.338130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f897700c-9c6e-4075-a65e-9301579332df {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.353425] env[62558]: DEBUG nova.compute.provider_tree [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.414835] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.586767] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266527, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.596993] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.706394] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Successfully created port: 2a1bff7c-65d5-4198-99c0-4566c13d38be {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.857833] env[62558]: DEBUG nova.scheduler.client.report [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.087251] env[62558]: DEBUG oslo_vmware.api [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266527, 'name': PowerOnVM_Task, 'duration_secs': 0.874847} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.087445] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 583.087646] env[62558]: DEBUG nova.compute.manager [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 583.088421] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6d4f6e-77c0-4c85-871d-3cd5de48e64d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.099213] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Releasing lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.099626] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.099812] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.100105] env[62558]: DEBUG oslo_concurrency.lockutils [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] Acquired lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.100300] env[62558]: DEBUG nova.network.neutron [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Refreshing network info cache for port c83e936a-60b1-40bd-ba75-3d7c01abc582 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.101231] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04c68339-4b6c-4f1c-b932-1a18f95dfd16 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.111392] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99292e11-b347-4712-860c-bd897df04117 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.142734] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5845d273-b18a-447a-ad2d-a011bfcc2423 could not be found. [ 583.142860] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.143044] env[62558]: INFO nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Took 0.04 seconds to destroy the instance on the hypervisor. [ 583.143301] env[62558]: DEBUG oslo.service.loopingcall [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.144168] env[62558]: DEBUG nova.compute.manager [-] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.145589] env[62558]: DEBUG nova.network.neutron [-] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.196247] env[62558]: DEBUG nova.network.neutron [-] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.397019] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 583.397589] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.099s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.399936] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.367s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.404911] env[62558]: INFO nova.compute.claims [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.433227] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 583.433227] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 583.433227] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 583.433381] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 583.433381] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 583.433448] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 583.435331] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 583.435331] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 583.435331] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 583.435331] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 583.435331] env[62558]: DEBUG nova.virt.hardware [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 583.435896] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3a85a8-f9b9-4f9a-a083-d0b92ba01f81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.439574] env[62558]: INFO nova.scheduler.client.report [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Deleted allocations for instance d02e6383-42d9-40db-89a3-ba8b5595b95b [ 583.449023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d7fda4-f306-40c9-ab0d-bd9295d09e82 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.605783] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.698587] env[62558]: DEBUG nova.network.neutron [-] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.797591] env[62558]: DEBUG nova.network.neutron [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.936873] env[62558]: DEBUG nova.network.neutron [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.950350] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d36b51fd-cc91-46ad-963e-e409989e40cc tempest-ServersAdmin275Test-1369229997 tempest-ServersAdmin275Test-1369229997-project-member] Lock "d02e6383-42d9-40db-89a3-ba8b5595b95b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.870s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.201765] env[62558]: INFO nova.compute.manager [-] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Took 1.06 seconds to deallocate network for instance. [ 584.205461] env[62558]: DEBUG nova.compute.claims [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.205636] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.440316] env[62558]: DEBUG oslo_concurrency.lockutils [req-9a91f315-7d57-45d1-b006-ece9db6ed2e3 req-cff1c61d-80fb-4d88-a20d-4663b025fca6 service nova] Releasing lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.732695] env[62558]: DEBUG nova.compute.manager [req-9ed126ec-0ac9-4c65-8e80-d650db77364b req-0a2dc832-fa6c-4b1b-8e47-7c18c532c18e service nova] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Received event network-vif-deleted-c83e936a-60b1-40bd-ba75-3d7c01abc582 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.937015] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2629cc9c-413c-4025-9aab-da4c8b0d1e33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.945647] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c3da5d-f833-4f80-a27f-21a89238c3c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.997804] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f426cbd8-df67-43cb-baf9-e8db308dab43 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.004446] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "9316facd-0c95-4720-b0d3-c0a39df81619" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.004446] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.004446] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "9316facd-0c95-4720-b0d3-c0a39df81619-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.004446] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.005062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.006962] env[62558]: INFO nova.compute.manager [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Terminating instance [ 585.011529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "refresh_cache-9316facd-0c95-4720-b0d3-c0a39df81619" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.011695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquired lock "refresh_cache-9316facd-0c95-4720-b0d3-c0a39df81619" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.011865] env[62558]: DEBUG nova.network.neutron [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.017020] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490f0381-46c7-43b7-bfa4-af6703043cd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.030578] env[62558]: DEBUG nova.compute.provider_tree [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.033363] env[62558]: ERROR nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 585.033363] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.033363] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.033363] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.033363] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.033363] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.033363] env[62558]: ERROR nova.compute.manager raise self.value [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.033363] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.033363] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.033363] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.034192] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.034192] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.034192] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 585.034192] env[62558]: ERROR nova.compute.manager [ 585.034192] env[62558]: Traceback (most recent call last): [ 585.034192] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.034192] env[62558]: listener.cb(fileno) [ 585.034192] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.034192] env[62558]: result = function(*args, **kwargs) [ 585.034192] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.034192] env[62558]: return func(*args, **kwargs) [ 585.034192] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.034192] env[62558]: raise e [ 585.034192] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.034192] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 585.034192] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.034192] env[62558]: created_port_ids = self._update_ports_for_instance( [ 585.034192] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.034192] env[62558]: with excutils.save_and_reraise_exception(): [ 585.034192] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.034192] env[62558]: self.force_reraise() [ 585.034192] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.034192] env[62558]: raise self.value [ 585.034192] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.034192] env[62558]: updated_port = self._update_port( [ 585.034192] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.034192] env[62558]: _ensure_no_port_binding_failure(port) [ 585.034192] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.034192] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.035103] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 585.035103] env[62558]: Removing descriptor: 18 [ 585.035103] env[62558]: ERROR nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Traceback (most recent call last): [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] yield resources [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.driver.spawn(context, instance, image_meta, [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.035103] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] vm_ref = self.build_virtual_machine(instance, [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] for vif in network_info: [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self._sync_wrapper(fn, *args, **kwargs) [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.wait() [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self[:] = self._gt.wait() [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self._exit_event.wait() [ 585.035467] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] result = hub.switch() [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self.greenlet.switch() [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] result = function(*args, **kwargs) [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return func(*args, **kwargs) [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise e [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] nwinfo = self.network_api.allocate_for_instance( [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.035862] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] created_port_ids = self._update_ports_for_instance( [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] with excutils.save_and_reraise_exception(): [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.force_reraise() [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise self.value [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] updated_port = self._update_port( [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] _ensure_no_port_binding_failure(port) [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.036288] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise exception.PortBindingFailed(port_id=port['id']) [ 585.036688] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 585.036688] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] [ 585.036688] env[62558]: INFO nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Terminating instance [ 585.036688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.036688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquired lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.036688] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.532254] env[62558]: DEBUG nova.scheduler.client.report [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.551496] env[62558]: DEBUG nova.network.neutron [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.561856] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.688766] env[62558]: DEBUG nova.network.neutron [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.748469] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.039984] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.640s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.040640] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 586.044479] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.936s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.046678] env[62558]: INFO nova.compute.claims [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.191788] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Releasing lock "refresh_cache-9316facd-0c95-4720-b0d3-c0a39df81619" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.192248] env[62558]: DEBUG nova.compute.manager [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.192455] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.193348] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd4c50c-5e9a-40ff-b477-52d4f9632885 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.203620] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 586.203620] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b0e80d2-9231-4732-9366-9fddc836b33b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.212063] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 586.212063] env[62558]: value = "task-1266530" [ 586.212063] env[62558]: _type = "Task" [ 586.212063] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.223191] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266530, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.251939] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Releasing lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.252461] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.252697] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.253031] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-533901e9-4ab3-4d78-9b27-b4d3d2b5e1d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.266700] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17931212-aec3-4ea0-aaea-9c69348aa71b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.292257] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f could not be found. [ 586.292522] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.292734] env[62558]: INFO nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 586.292996] env[62558]: DEBUG oslo.service.loopingcall [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.293252] env[62558]: DEBUG nova.compute.manager [-] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.293351] env[62558]: DEBUG nova.network.neutron [-] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.312485] env[62558]: DEBUG nova.network.neutron [-] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.550907] env[62558]: DEBUG nova.compute.utils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.559725] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 586.723345] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266530, 'name': PowerOffVM_Task, 'duration_secs': 0.12527} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.723725] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 586.724008] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 586.724852] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46ed3a5e-c622-4ba6-a4ae-eaddbdf9728a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.755944] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 586.755944] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 586.756378] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Deleting the datastore file [datastore2] 9316facd-0c95-4720-b0d3-c0a39df81619 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 586.756457] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8a27c9d-89c8-4ec2-846a-dcc7ebb9b3a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.764235] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for the task: (returnval){ [ 586.764235] env[62558]: value = "task-1266532" [ 586.764235] env[62558]: _type = "Task" [ 586.764235] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.775752] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.817530] env[62558]: DEBUG nova.network.neutron [-] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.865240] env[62558]: DEBUG nova.compute.manager [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Received event network-changed-2a1bff7c-65d5-4198-99c0-4566c13d38be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.865240] env[62558]: DEBUG nova.compute.manager [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Refreshing instance network info cache due to event network-changed-2a1bff7c-65d5-4198-99c0-4566c13d38be. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.865448] env[62558]: DEBUG oslo_concurrency.lockutils [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] Acquiring lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.865562] env[62558]: DEBUG oslo_concurrency.lockutils [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] Acquired lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.865769] env[62558]: DEBUG nova.network.neutron [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Refreshing network info cache for port 2a1bff7c-65d5-4198-99c0-4566c13d38be {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.065494] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 587.274904] env[62558]: DEBUG oslo_vmware.api [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Task: {'id': task-1266532, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094885} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.277775] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.277970] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 587.278166] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.278339] env[62558]: INFO nova.compute.manager [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Took 1.09 seconds to destroy the instance on the hypervisor. [ 587.278580] env[62558]: DEBUG oslo.service.loopingcall [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.279450] env[62558]: DEBUG nova.compute.manager [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.279776] env[62558]: DEBUG nova.network.neutron [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.301278] env[62558]: DEBUG nova.network.neutron [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.321196] env[62558]: INFO nova.compute.manager [-] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Took 1.03 seconds to deallocate network for instance. [ 587.324298] env[62558]: DEBUG nova.compute.claims [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.324391] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.402407] env[62558]: DEBUG nova.network.neutron [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.527355] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a5795a-ce14-4764-9d2e-13b1f830b19f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.541335] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893889fc-a99f-47f6-8c51-487c8a0571ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.582732] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cf27d5-9839-4674-a72a-6f8e7a7b6468 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.587932] env[62558]: DEBUG nova.network.neutron [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.591757] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84204bac-233b-40c5-b2e3-64e83b4f41a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.607489] env[62558]: DEBUG nova.compute.provider_tree [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.804175] env[62558]: DEBUG nova.network.neutron [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.088300] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 588.092562] env[62558]: DEBUG oslo_concurrency.lockutils [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] Releasing lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.092562] env[62558]: DEBUG nova.compute.manager [req-a672594e-d797-45c1-bced-5cc17572af97 req-bd110edf-6b66-4590-96a9-9f1df3ad10f9 service nova] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Received event network-vif-deleted-2a1bff7c-65d5-4198-99c0-4566c13d38be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.113130] env[62558]: DEBUG nova.scheduler.client.report [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.127008] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 588.127287] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 588.127403] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.127573] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 588.127709] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.128116] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 588.128116] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 588.128356] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 588.128356] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 588.128869] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 588.128869] env[62558]: DEBUG nova.virt.hardware [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 588.129582] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f45e42a-082f-4240-a0ff-3ad993847809 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.140392] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917aed08-1936-486c-83df-f96a7ff78e08 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.158238] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 588.164196] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Creating folder: Project (19e15f21b11341ff9c22912ed08edfef). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.164534] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bcac0cc-ff7c-4b78-bb1b-3e9a3b29d54c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.175283] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Created folder: Project (19e15f21b11341ff9c22912ed08edfef) in parent group-v272451. [ 588.175519] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Creating folder: Instances. Parent ref: group-v272468. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 588.175751] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fff887b6-1b22-4296-bc63-8f1fea7690c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.184563] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Created folder: Instances in parent group-v272468. [ 588.184880] env[62558]: DEBUG oslo.service.loopingcall [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.185189] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 588.185478] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4714332-7183-459a-937c-6591db31b81a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.203912] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 588.203912] env[62558]: value = "task-1266536" [ 588.203912] env[62558]: _type = "Task" [ 588.203912] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.212287] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266536, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.307365] env[62558]: INFO nova.compute.manager [-] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Took 1.03 seconds to deallocate network for instance. [ 588.633898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.633898] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.633898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.578s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.633898] env[62558]: INFO nova.compute.claims [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.719067] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266536, 'name': CreateVM_Task, 'duration_secs': 0.252861} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.719468] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 588.720062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.720605] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.720720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 588.721368] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af5eada8-e2d6-455d-a40b-d15ce54e8ee1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.728415] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 588.728415] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523221bb-dab7-427a-30ef-ec27794d86ac" [ 588.728415] env[62558]: _type = "Task" [ 588.728415] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.739331] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523221bb-dab7-427a-30ef-ec27794d86ac, 'name': SearchDatastore_Task, 'duration_secs': 0.009143} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.739747] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.740084] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 588.740741] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.740999] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.741273] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.741659] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ff3c9e7-0c16-4987-96e9-57d9883e75d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.751956] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.752248] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 588.753125] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0784c5f-25de-4221-a2d8-2c58bc220c2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.766014] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 588.766014] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52dcd0fc-8adb-ebfe-05af-210472548e4b" [ 588.766014] env[62558]: _type = "Task" [ 588.766014] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.780564] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52dcd0fc-8adb-ebfe-05af-210472548e4b, 'name': SearchDatastore_Task, 'duration_secs': 0.008737} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.782108] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76dd81d3-13e5-4316-baf7-270b96a85346 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.788066] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 588.788066] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528c638a-c19c-b372-09dc-2137d4f722b7" [ 588.788066] env[62558]: _type = "Task" [ 588.788066] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.797055] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528c638a-c19c-b372-09dc-2137d4f722b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.815340] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.135566] env[62558]: DEBUG nova.compute.utils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.139829] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.139829] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 589.215428] env[62558]: DEBUG nova.policy [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08837d0811c949b4871d0c5a50b73229', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '440c42b1b44a4373aca7068989e23513', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.301526] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528c638a-c19c-b372-09dc-2137d4f722b7, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.301526] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.301526] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 589.301526] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da2da6ff-0489-405b-9953-e2dcc3a90893 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.306139] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 589.306139] env[62558]: value = "task-1266537" [ 589.306139] env[62558]: _type = "Task" [ 589.306139] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.314824] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.648159] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.707040] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Successfully created port: 1d437e03-f755-47ff-9d65-0951fcbd5722 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.818064] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266537, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463395} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.820721] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 589.820931] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 589.821418] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d4c326f-82ba-414b-9513-237436935ce1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.828222] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 589.828222] env[62558]: value = "task-1266538" [ 589.828222] env[62558]: _type = "Task" [ 589.828222] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.836091] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266538, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.099607] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8173fb73-856b-4700-ad3b-46282d9c49dd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.106379] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d69e25-1d6d-4f48-9975-c3cb62328920 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.140853] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c184df60-379f-4644-9f0c-fe072bd02591 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.148184] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe0184b-6a25-47b0-bce4-470669d90cdd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.166886] env[62558]: DEBUG nova.compute.provider_tree [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.338706] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266538, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065084} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.339094] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 590.339976] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f441418-08c6-496f-91aa-b907201cf2d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.362043] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 590.362043] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e22b47ce-9758-4815-bc58-bae3409ecdd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.382729] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 590.382729] env[62558]: value = "task-1266539" [ 590.382729] env[62558]: _type = "Task" [ 590.382729] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.391838] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266539, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.661213] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.671972] env[62558]: DEBUG nova.scheduler.client.report [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.690306] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.690565] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.690717] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.691582] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.691756] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.691906] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.693303] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.693303] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.693303] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.693303] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.693303] env[62558]: DEBUG nova.virt.hardware [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.693692] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed76131d-fa58-4501-9bad-80b75e00174a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.703657] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81959965-dcfe-4c8b-be8c-da2f5b76948f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.893396] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266539, 'name': ReconfigVM_Task, 'duration_secs': 0.251725} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.893396] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Reconfigured VM instance instance-00000012 to attach disk [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 590.893587] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-890eba74-af16-410a-8985-ed1e952622f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.900625] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 590.900625] env[62558]: value = "task-1266540" [ 590.900625] env[62558]: _type = "Task" [ 590.900625] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.909771] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266540, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.181803] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.182378] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.188359] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.535s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.188359] env[62558]: INFO nova.compute.claims [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.391234] env[62558]: ERROR nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 591.391234] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.391234] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.391234] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.391234] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.391234] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.391234] env[62558]: ERROR nova.compute.manager raise self.value [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.391234] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.391234] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.391234] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.391658] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.391658] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.391658] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 591.391658] env[62558]: ERROR nova.compute.manager [ 591.391658] env[62558]: Traceback (most recent call last): [ 591.391658] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.391658] env[62558]: listener.cb(fileno) [ 591.391658] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.391658] env[62558]: result = function(*args, **kwargs) [ 591.391658] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.391658] env[62558]: return func(*args, **kwargs) [ 591.391658] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.391658] env[62558]: raise e [ 591.391658] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.391658] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 591.391658] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.391658] env[62558]: created_port_ids = self._update_ports_for_instance( [ 591.391658] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.391658] env[62558]: with excutils.save_and_reraise_exception(): [ 591.391658] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.391658] env[62558]: self.force_reraise() [ 591.391658] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.391658] env[62558]: raise self.value [ 591.391658] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.391658] env[62558]: updated_port = self._update_port( [ 591.391658] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.391658] env[62558]: _ensure_no_port_binding_failure(port) [ 591.391658] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.391658] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.392385] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 591.392385] env[62558]: Removing descriptor: 22 [ 591.392385] env[62558]: ERROR nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Traceback (most recent call last): [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] yield resources [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.driver.spawn(context, instance, image_meta, [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.392385] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] vm_ref = self.build_virtual_machine(instance, [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] for vif in network_info: [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self._sync_wrapper(fn, *args, **kwargs) [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.wait() [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self[:] = self._gt.wait() [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self._exit_event.wait() [ 591.392664] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] result = hub.switch() [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self.greenlet.switch() [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] result = function(*args, **kwargs) [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return func(*args, **kwargs) [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise e [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] nwinfo = self.network_api.allocate_for_instance( [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.392969] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] created_port_ids = self._update_ports_for_instance( [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] with excutils.save_and_reraise_exception(): [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.force_reraise() [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise self.value [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] updated_port = self._update_port( [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] _ensure_no_port_binding_failure(port) [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.393289] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise exception.PortBindingFailed(port_id=port['id']) [ 591.393685] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 591.393685] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] [ 591.393685] env[62558]: INFO nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Terminating instance [ 591.394652] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquiring lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.394808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquired lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.395109] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.409963] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266540, 'name': Rename_Task, 'duration_secs': 0.148539} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.411618] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 591.411868] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3305ee73-f92b-428b-ad08-d9fb060e9608 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.418915] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 591.418915] env[62558]: value = "task-1266541" [ 591.418915] env[62558]: _type = "Task" [ 591.418915] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.430444] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.694217] env[62558]: DEBUG nova.compute.utils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.695817] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.696117] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.750144] env[62558]: DEBUG nova.policy [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbd32f833a88422f850c86d764de3736', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81ebf6dfdbcd46098ad0e978899cd27b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.810141] env[62558]: DEBUG nova.compute.manager [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Received event network-changed-1d437e03-f755-47ff-9d65-0951fcbd5722 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.810141] env[62558]: DEBUG nova.compute.manager [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Refreshing instance network info cache due to event network-changed-1d437e03-f755-47ff-9d65-0951fcbd5722. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.810141] env[62558]: DEBUG oslo_concurrency.lockutils [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] Acquiring lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.927718] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.940743] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266541, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.148933] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.200304] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquiring lock "2ae701c0-3016-4635-8fe1-1e1c314fa748" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.200572] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "2ae701c0-3016-4635-8fe1-1e1c314fa748" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.203634] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.430064] env[62558]: DEBUG oslo_vmware.api [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266541, 'name': PowerOnVM_Task, 'duration_secs': 0.667735} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.430673] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 592.430673] env[62558]: INFO nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Took 4.34 seconds to spawn the instance on the hypervisor. [ 592.430880] env[62558]: DEBUG nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 592.431922] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e4379c-052f-4dd6-98d7-49a9817c224b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.591088] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Successfully created port: 5d42bc9c-0db8-4264-a494-384dcc6f36dc {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.655030] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Releasing lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.655030] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.655030] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.655030] env[62558]: DEBUG oslo_concurrency.lockutils [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] Acquired lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.655030] env[62558]: DEBUG nova.network.neutron [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Refreshing network info cache for port 1d437e03-f755-47ff-9d65-0951fcbd5722 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.655327] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2749e60d-319d-45a2-b6c4-a705b76b4c8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.668227] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeacb172-543d-4b23-be4b-fd21cf8d0ab4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.698512] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 813e287c-9d7f-4e9c-9216-986f631b6eb6 could not be found. [ 592.698512] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 592.698512] env[62558]: INFO nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.698681] env[62558]: DEBUG oslo.service.loopingcall [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.701928] env[62558]: DEBUG nova.compute.manager [-] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.702079] env[62558]: DEBUG nova.network.neutron [-] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.744657] env[62558]: DEBUG nova.network.neutron [-] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.794517] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d74cc98-6bdd-4eb1-8f08-b958f3a5108d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.803076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2f6754-1bdd-441c-996e-67629b4dae65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.835053] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae0fb1f-e003-44f0-a69c-bcb375d39450 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.843183] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffddc95c-bef4-41d4-a292-74ee51773836 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.858319] env[62558]: DEBUG nova.compute.provider_tree [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.951958] env[62558]: INFO nova.compute.manager [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Took 29.94 seconds to build instance. [ 593.220518] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.247872] env[62558]: DEBUG nova.network.neutron [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.249347] env[62558]: DEBUG nova.network.neutron [-] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.253015] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.253604] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.253604] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.254344] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.254344] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.254344] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.255626] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.256219] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.256219] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.256219] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.256462] env[62558]: DEBUG nova.virt.hardware [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.257486] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c99a4f-708e-4ce3-9b71-4dcf1b30e400 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.269275] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e390df9-6ac8-4e77-8b30-e618293d1b69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.366021] env[62558]: DEBUG nova.scheduler.client.report [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.453640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f5da14b-b76b-4d07-8caf-59b636489f35 tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.082s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.517311] env[62558]: DEBUG nova.network.neutron [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.763196] env[62558]: INFO nova.compute.manager [-] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Took 1.06 seconds to deallocate network for instance. [ 593.764225] env[62558]: DEBUG nova.compute.claims [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.765510] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.877594] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.692s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.878194] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.882784] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.360s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.884351] env[62558]: INFO nova.compute.claims [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.957330] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 594.023018] env[62558]: DEBUG oslo_concurrency.lockutils [req-a2b0a96b-79a0-4081-a5fe-e5792fd51f8b req-39a15901-58dd-46f1-bc2d-5a9b917da379 service nova] Releasing lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.183325] env[62558]: DEBUG nova.compute.manager [req-f298f676-97a4-429a-ba61-fd94c8fc8f6c req-cc63e5f1-4119-4397-bd73-e22f7d12baca service nova] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Received event network-vif-deleted-1d437e03-f755-47ff-9d65-0951fcbd5722 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.392898] env[62558]: DEBUG nova.compute.utils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.400025] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 594.403018] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.477451] env[62558]: DEBUG nova.policy [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8801c599f26a4eaaae9051ddb8a247ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2ee76d3ce9f44c88762335678dc8e46', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.483409] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.822795] env[62558]: ERROR nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 594.822795] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.822795] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.822795] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.822795] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.822795] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.822795] env[62558]: ERROR nova.compute.manager raise self.value [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.822795] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.822795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.822795] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.823471] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.823471] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.823471] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 594.823471] env[62558]: ERROR nova.compute.manager [ 594.823471] env[62558]: Traceback (most recent call last): [ 594.823471] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.823471] env[62558]: listener.cb(fileno) [ 594.823471] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.823471] env[62558]: result = function(*args, **kwargs) [ 594.823471] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.823471] env[62558]: return func(*args, **kwargs) [ 594.823471] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.823471] env[62558]: raise e [ 594.823471] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.823471] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 594.823471] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.823471] env[62558]: created_port_ids = self._update_ports_for_instance( [ 594.823471] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.823471] env[62558]: with excutils.save_and_reraise_exception(): [ 594.823471] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.823471] env[62558]: self.force_reraise() [ 594.823471] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.823471] env[62558]: raise self.value [ 594.823471] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.823471] env[62558]: updated_port = self._update_port( [ 594.823471] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.823471] env[62558]: _ensure_no_port_binding_failure(port) [ 594.823471] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.823471] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.824153] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 594.824153] env[62558]: Removing descriptor: 22 [ 594.824153] env[62558]: ERROR nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Traceback (most recent call last): [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] yield resources [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.driver.spawn(context, instance, image_meta, [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.824153] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] vm_ref = self.build_virtual_machine(instance, [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] for vif in network_info: [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self._sync_wrapper(fn, *args, **kwargs) [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.wait() [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self[:] = self._gt.wait() [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self._exit_event.wait() [ 594.824567] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] result = hub.switch() [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self.greenlet.switch() [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] result = function(*args, **kwargs) [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return func(*args, **kwargs) [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise e [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] nwinfo = self.network_api.allocate_for_instance( [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.824895] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] created_port_ids = self._update_ports_for_instance( [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] with excutils.save_and_reraise_exception(): [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.force_reraise() [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise self.value [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] updated_port = self._update_port( [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] _ensure_no_port_binding_failure(port) [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.829562] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise exception.PortBindingFailed(port_id=port['id']) [ 594.829941] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 594.829941] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] [ 594.829941] env[62558]: INFO nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Terminating instance [ 594.829941] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.829941] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquired lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.829941] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.900524] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.069060] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Successfully created port: c85a916b-8b11-45f3-ac31-b803326f3461 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.342283] env[62558]: INFO nova.compute.manager [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Rebuilding instance [ 595.358326] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.391630] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17898fce-cb21-4016-a513-d4eb3419d51d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.396712] env[62558]: DEBUG nova.compute.manager [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 595.397476] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d312a01-84a8-4b97-bcdb-58ef6b828086 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.403179] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc7b0f7-54b0-41f1-93bb-e2b3fd874ea7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.447741] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798572d1-bed2-4646-afc3-95f1f4a49fbd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.456549] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1ce2e6-48c9-47cc-81e1-fa7941e796c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.472694] env[62558]: DEBUG nova.compute.provider_tree [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.526512] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.917059] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.920507] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 595.921083] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-802cacff-d83f-4d9e-9aec-80b0ea672ea5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.930119] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 595.930119] env[62558]: value = "task-1266542" [ 595.930119] env[62558]: _type = "Task" [ 595.930119] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.940787] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.943104] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.943359] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.943512] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.943692] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.943909] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.943990] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.944194] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.944342] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.944501] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.944659] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.944910] env[62558]: DEBUG nova.virt.hardware [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.946343] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6da5a37-01d3-4d4c-9e90-a0ca1bf78bf4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.958450] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9e4d7e-4923-4e02-810e-07adfbbd519e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.975489] env[62558]: DEBUG nova.scheduler.client.report [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.032021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Releasing lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.032021] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.032021] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.032021] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54764bf6-2291-4777-bbc8-84eae1893af8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.040614] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c5f676-faaf-4ff0-b125-6dab1e2e60c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.064039] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e could not be found. [ 596.064422] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 596.064732] env[62558]: INFO nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 596.065163] env[62558]: DEBUG oslo.service.loopingcall [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.065526] env[62558]: DEBUG nova.compute.manager [-] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.065741] env[62558]: DEBUG nova.network.neutron [-] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.087032] env[62558]: DEBUG nova.network.neutron [-] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.339177] env[62558]: DEBUG nova.compute.manager [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Received event network-changed-5d42bc9c-0db8-4264-a494-384dcc6f36dc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.339376] env[62558]: DEBUG nova.compute.manager [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Refreshing instance network info cache due to event network-changed-5d42bc9c-0db8-4264-a494-384dcc6f36dc. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.339585] env[62558]: DEBUG oslo_concurrency.lockutils [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] Acquiring lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.339927] env[62558]: DEBUG oslo_concurrency.lockutils [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] Acquired lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.339927] env[62558]: DEBUG nova.network.neutron [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Refreshing network info cache for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 596.445165] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266542, 'name': PowerOffVM_Task, 'duration_secs': 0.206743} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.445478] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 596.445720] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.447012] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a3827a-1ef1-40d9-a1fa-fb569006bb9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.453654] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 596.453906] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ae0091f-2bda-4e1a-bfe7-72031c331c49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.478622] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 596.478825] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 596.478995] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Deleting the datastore file [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 596.479265] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-824d9a68-4348-4ac3-870f-4ef5eae2597f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.487194] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.487578] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.491928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.530s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.501031] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 596.501031] env[62558]: value = "task-1266544" [ 596.501031] env[62558]: _type = "Task" [ 596.501031] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.511429] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266544, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.523361] env[62558]: ERROR nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 596.523361] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.523361] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.523361] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.523361] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.523361] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.523361] env[62558]: ERROR nova.compute.manager raise self.value [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.523361] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.523361] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.523361] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.524479] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.524479] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.524479] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 596.524479] env[62558]: ERROR nova.compute.manager [ 596.524479] env[62558]: Traceback (most recent call last): [ 596.524479] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.524479] env[62558]: listener.cb(fileno) [ 596.524479] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.524479] env[62558]: result = function(*args, **kwargs) [ 596.524479] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.524479] env[62558]: return func(*args, **kwargs) [ 596.524479] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.524479] env[62558]: raise e [ 596.524479] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.524479] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 596.524479] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.524479] env[62558]: created_port_ids = self._update_ports_for_instance( [ 596.524479] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.524479] env[62558]: with excutils.save_and_reraise_exception(): [ 596.524479] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.524479] env[62558]: self.force_reraise() [ 596.524479] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.524479] env[62558]: raise self.value [ 596.524479] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.524479] env[62558]: updated_port = self._update_port( [ 596.524479] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.524479] env[62558]: _ensure_no_port_binding_failure(port) [ 596.524479] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.524479] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.525289] env[62558]: nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 596.525289] env[62558]: Removing descriptor: 18 [ 596.525289] env[62558]: ERROR nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Traceback (most recent call last): [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] yield resources [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.driver.spawn(context, instance, image_meta, [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.525289] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] vm_ref = self.build_virtual_machine(instance, [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] for vif in network_info: [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self._sync_wrapper(fn, *args, **kwargs) [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.wait() [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self[:] = self._gt.wait() [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self._exit_event.wait() [ 596.525595] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] result = hub.switch() [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self.greenlet.switch() [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] result = function(*args, **kwargs) [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return func(*args, **kwargs) [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise e [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] nwinfo = self.network_api.allocate_for_instance( [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.525922] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] created_port_ids = self._update_ports_for_instance( [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] with excutils.save_and_reraise_exception(): [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.force_reraise() [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise self.value [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] updated_port = self._update_port( [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] _ensure_no_port_binding_failure(port) [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.526216] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise exception.PortBindingFailed(port_id=port['id']) [ 596.526487] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 596.526487] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] [ 596.526487] env[62558]: INFO nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Terminating instance [ 596.527308] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquiring lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.527526] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquired lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.528201] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.589936] env[62558]: DEBUG nova.network.neutron [-] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.899708] env[62558]: DEBUG nova.network.neutron [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.994100] env[62558]: DEBUG nova.compute.utils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.995509] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.995671] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 597.019690] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266544, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09281} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.019957] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 597.021340] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 597.021447] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.051198] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.093741] env[62558]: DEBUG nova.network.neutron [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.095806] env[62558]: INFO nova.compute.manager [-] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Took 1.03 seconds to deallocate network for instance. [ 597.099056] env[62558]: DEBUG nova.compute.claims [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.099109] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.113884] env[62558]: DEBUG nova.policy [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4cbe0033a0bb48d1ab4d03546d2d3493', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c516a8aef1614886b3b28d81167e81e0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.202999] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.499961] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e9edf5-cbe3-4ed3-a7c8-8b32ea472770 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.503122] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.511024] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c24d21-875e-48b2-a710-00f6cb6e904b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.546032] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a3c82e-3f9d-4c8d-a5ad-c4ae1aa35b61 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.555497] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c073015d-fc7e-4a93-9729-ecfa50cd40f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.573849] env[62558]: DEBUG nova.compute.provider_tree [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.599756] env[62558]: DEBUG oslo_concurrency.lockutils [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] Releasing lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.599756] env[62558]: DEBUG nova.compute.manager [req-b5c0c1cf-ce13-4d3c-a0ef-38098d4048db req-cb7275fe-ba2e-43a8-9b63-bf3a7dc5bf62 service nova] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Received event network-vif-deleted-5d42bc9c-0db8-4264-a494-384dcc6f36dc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.712113] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Releasing lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.712382] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 597.713942] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.714970] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d1e5863-a05e-4e87-a798-566a08e84cbd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.727042] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b828c8a-ac1f-4200-bcf1-4795061cad27 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.741884] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Successfully created port: 053e1b3b-2d48-4717-8475-9241c88ba08e {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.749952] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5a1f074f-8e43-42f9-864d-976cdaa4639c could not be found. [ 597.750302] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.750381] env[62558]: INFO nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 597.750656] env[62558]: DEBUG oslo.service.loopingcall [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.750830] env[62558]: DEBUG nova.compute.manager [-] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.750919] env[62558]: DEBUG nova.network.neutron [-] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.767627] env[62558]: DEBUG nova.network.neutron [-] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.077704] env[62558]: DEBUG nova.scheduler.client.report [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.083717] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.084216] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.084468] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.084747] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.085064] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.085310] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.089323] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.089504] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.089679] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.089854] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.090042] env[62558]: DEBUG nova.virt.hardware [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.091239] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b2b790-951d-4767-8aa0-314e03f83679 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.096744] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.605s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.098385] env[62558]: ERROR nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Traceback (most recent call last): [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.driver.spawn(context, instance, image_meta, [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] vm_ref = self.build_virtual_machine(instance, [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.098385] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] for vif in network_info: [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self._sync_wrapper(fn, *args, **kwargs) [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.wait() [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self[:] = self._gt.wait() [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self._exit_event.wait() [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] result = hub.switch() [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.098872] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return self.greenlet.switch() [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] result = function(*args, **kwargs) [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] return func(*args, **kwargs) [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise e [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] nwinfo = self.network_api.allocate_for_instance( [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] created_port_ids = self._update_ports_for_instance( [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] with excutils.save_and_reraise_exception(): [ 598.099367] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] self.force_reraise() [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise self.value [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] updated_port = self._update_port( [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] _ensure_no_port_binding_failure(port) [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] raise exception.PortBindingFailed(port_id=port['id']) [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] nova.exception.PortBindingFailed: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. [ 598.099818] env[62558]: ERROR nova.compute.manager [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] [ 598.100238] env[62558]: DEBUG nova.compute.utils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.100238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.572s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.101478] env[62558]: INFO nova.compute.claims [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.104062] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Build of instance f52921c6-abbc-4a27-a451-e1cb4f5dfc48 was re-scheduled: Binding failed for port 8b26b7d8-1344-489e-90c0-b07a7bc7ba89, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.104549] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.104755] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquiring lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.105114] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Acquired lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.105114] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.112024] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fe3158-85c6-464e-8ad9-566ac7497a4e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.126957] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 598.132427] env[62558]: DEBUG oslo.service.loopingcall [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.133243] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 598.133243] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a180d1c3-482f-44bd-a371-abf64212753d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.151610] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 598.151610] env[62558]: value = "task-1266545" [ 598.151610] env[62558]: _type = "Task" [ 598.151610] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.160725] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266545, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.270080] env[62558]: DEBUG nova.network.neutron [-] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.526028] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.558586] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.558888] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.558982] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.561383] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.561383] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.561383] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.561383] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.561383] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.561805] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.561805] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.561805] env[62558]: DEBUG nova.virt.hardware [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.562302] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10741adc-184a-4d7a-862d-e3902889d043 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.571432] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2377309-df7c-40bd-9e2f-57680e83b482 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.630742] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.667752] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266545, 'name': CreateVM_Task, 'duration_secs': 0.31582} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.667950] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 598.668409] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.668564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.668890] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 598.669174] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aba7712-ae7c-4311-a92a-1cf87b8833cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.674395] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 598.674395] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524c5bc5-93ca-f392-6a45-47598ca451ba" [ 598.674395] env[62558]: _type = "Task" [ 598.674395] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.685124] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524c5bc5-93ca-f392-6a45-47598ca451ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.730055] env[62558]: DEBUG nova.compute.manager [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Received event network-changed-c85a916b-8b11-45f3-ac31-b803326f3461 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.730259] env[62558]: DEBUG nova.compute.manager [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Refreshing instance network info cache due to event network-changed-c85a916b-8b11-45f3-ac31-b803326f3461. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 598.730475] env[62558]: DEBUG oslo_concurrency.lockutils [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] Acquiring lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.730615] env[62558]: DEBUG oslo_concurrency.lockutils [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] Acquired lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.730763] env[62558]: DEBUG nova.network.neutron [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Refreshing network info cache for port c85a916b-8b11-45f3-ac31-b803326f3461 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.762222] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.772655] env[62558]: INFO nova.compute.manager [-] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Took 1.02 seconds to deallocate network for instance. [ 598.775744] env[62558]: DEBUG nova.compute.claims [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.775954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.188297] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524c5bc5-93ca-f392-6a45-47598ca451ba, 'name': SearchDatastore_Task, 'duration_secs': 0.014161} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.188663] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.189140] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 599.189427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.189541] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.189716] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 599.190060] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0aafe88c-99dc-4070-a587-abe77b0951d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.205099] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 599.205099] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 599.205099] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb92044-a8fa-4040-aa8b-dcc69ab1e28d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.209605] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 599.209605] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5257c2fc-90f1-9e9e-88f9-fbd635736e90" [ 599.209605] env[62558]: _type = "Task" [ 599.209605] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.221140] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5257c2fc-90f1-9e9e-88f9-fbd635736e90, 'name': SearchDatastore_Task, 'duration_secs': 0.008835} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.224802] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d6692cb-a7e9-4d04-9b94-55831d7a1e7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.231490] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 599.231490] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524f0da8-b43e-598e-41fb-e2a8965240ce" [ 599.231490] env[62558]: _type = "Task" [ 599.231490] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.247561] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524f0da8-b43e-598e-41fb-e2a8965240ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011116} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.248826] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.248826] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 599.249087] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1708af7b-c485-456d-9a23-7c54ad02b5eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.256356] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 599.256356] env[62558]: value = "task-1266546" [ 599.256356] env[62558]: _type = "Task" [ 599.256356] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.261285] env[62558]: DEBUG nova.network.neutron [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.266342] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Releasing lock "refresh_cache-f52921c6-abbc-4a27-a451-e1cb4f5dfc48" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.266616] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.268361] env[62558]: DEBUG nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.268361] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.272525] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.298965] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.373766] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquiring lock "cecebc6a-25cc-4376-8467-2aa0f909c9bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.374190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "cecebc6a-25cc-4376-8467-2aa0f909c9bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.384381] env[62558]: DEBUG nova.network.neutron [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.595410] env[62558]: ERROR nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 599.595410] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.595410] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.595410] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.595410] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.595410] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.595410] env[62558]: ERROR nova.compute.manager raise self.value [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.595410] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.595410] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.595410] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.596038] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.596038] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.596038] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 599.596038] env[62558]: ERROR nova.compute.manager [ 599.596038] env[62558]: Traceback (most recent call last): [ 599.596038] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.596038] env[62558]: listener.cb(fileno) [ 599.596038] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.596038] env[62558]: result = function(*args, **kwargs) [ 599.596038] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.596038] env[62558]: return func(*args, **kwargs) [ 599.596038] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.596038] env[62558]: raise e [ 599.596038] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.596038] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 599.596038] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.596038] env[62558]: created_port_ids = self._update_ports_for_instance( [ 599.596038] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.596038] env[62558]: with excutils.save_and_reraise_exception(): [ 599.596038] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.596038] env[62558]: self.force_reraise() [ 599.596038] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.596038] env[62558]: raise self.value [ 599.596038] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.596038] env[62558]: updated_port = self._update_port( [ 599.596038] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.596038] env[62558]: _ensure_no_port_binding_failure(port) [ 599.596038] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.596038] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.596738] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 599.596738] env[62558]: Removing descriptor: 22 [ 599.596738] env[62558]: ERROR nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Traceback (most recent call last): [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] yield resources [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.driver.spawn(context, instance, image_meta, [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.596738] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] vm_ref = self.build_virtual_machine(instance, [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] for vif in network_info: [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self._sync_wrapper(fn, *args, **kwargs) [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.wait() [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self[:] = self._gt.wait() [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self._exit_event.wait() [ 599.597041] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] result = hub.switch() [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self.greenlet.switch() [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] result = function(*args, **kwargs) [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return func(*args, **kwargs) [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise e [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] nwinfo = self.network_api.allocate_for_instance( [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.597396] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] created_port_ids = self._update_ports_for_instance( [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] with excutils.save_and_reraise_exception(): [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.force_reraise() [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise self.value [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] updated_port = self._update_port( [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] _ensure_no_port_binding_failure(port) [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.597725] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise exception.PortBindingFailed(port_id=port['id']) [ 599.598033] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 599.598033] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] [ 599.598033] env[62558]: INFO nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Terminating instance [ 599.598033] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquiring lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.598033] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquired lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.598033] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.649271] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1c5037-a18f-41a6-8957-aebdc4d6a1d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.659837] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d979476-35c4-4048-b742-de30c09a227d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.700936] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38db737e-c9dd-4922-b1d3-a66d36aec7c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.710011] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d21175c-e9d8-4e6d-9541-760f2d34b0f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.723290] env[62558]: DEBUG nova.compute.provider_tree [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.767792] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465707} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.767910] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.768404] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 599.768701] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1935fddf-dc76-487a-bd18-151d10cbbe5d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.775423] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 599.775423] env[62558]: value = "task-1266547" [ 599.775423] env[62558]: _type = "Task" [ 599.775423] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.784576] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.803021] env[62558]: DEBUG nova.network.neutron [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.886694] env[62558]: DEBUG oslo_concurrency.lockutils [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] Releasing lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.886975] env[62558]: DEBUG nova.compute.manager [req-ca5497b1-cbc0-464d-895d-5a5c09f7a5ef req-03c818f2-1e86-470d-9a88-24e35a24851d service nova] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Received event network-vif-deleted-c85a916b-8b11-45f3-ac31-b803326f3461 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.120195] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.227204] env[62558]: DEBUG nova.scheduler.client.report [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.252746] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.288141] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057923} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.288379] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 600.289567] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b393f9f-4021-4d60-98d8-e88f75e9694e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.310575] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 600.311111] env[62558]: INFO nova.compute.manager [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] [instance: f52921c6-abbc-4a27-a451-e1cb4f5dfc48] Took 1.04 seconds to deallocate network for instance. [ 600.314184] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52759b2f-6214-436d-809a-e2722f64a738 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.333993] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 600.333993] env[62558]: value = "task-1266548" [ 600.333993] env[62558]: _type = "Task" [ 600.333993] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.341964] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266548, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.558434] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquiring lock "a092fb50-c750-4b91-a94e-e11e9eb7faf5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.558688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "a092fb50-c750-4b91-a94e-e11e9eb7faf5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.731759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.732430] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.736086] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.990s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.755405] env[62558]: DEBUG nova.compute.manager [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Received event network-changed-053e1b3b-2d48-4717-8475-9241c88ba08e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.756058] env[62558]: DEBUG nova.compute.manager [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Refreshing instance network info cache due to event network-changed-053e1b3b-2d48-4717-8475-9241c88ba08e. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.756285] env[62558]: DEBUG oslo_concurrency.lockutils [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] Acquiring lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.756653] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Releasing lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.757019] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 600.757341] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 600.757890] env[62558]: DEBUG oslo_concurrency.lockutils [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] Acquired lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.758071] env[62558]: DEBUG nova.network.neutron [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Refreshing network info cache for port 053e1b3b-2d48-4717-8475-9241c88ba08e {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 600.759423] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91f5a67e-1865-416a-8451-ca4b37ac1ab4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.769876] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7089768a-199f-4f26-979d-f7e4a615e86c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.796975] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1513843d-0a96-4b56-86bb-3c39ab4b3692 could not be found. [ 600.797293] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 600.797537] env[62558]: INFO nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Took 0.04 seconds to destroy the instance on the hypervisor. [ 600.797864] env[62558]: DEBUG oslo.service.loopingcall [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.798195] env[62558]: DEBUG nova.compute.manager [-] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.798527] env[62558]: DEBUG nova.network.neutron [-] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.814161] env[62558]: DEBUG nova.network.neutron [-] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.845574] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266548, 'name': ReconfigVM_Task, 'duration_secs': 0.254879} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.845909] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Reconfigured VM instance instance-00000012 to attach disk [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151/f4f7e2c9-229c-4547-b674-5a5219a28151.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.846555] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38b9e3ee-fea8-4cc5-ac53-8281efb3b9f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.852868] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 600.852868] env[62558]: value = "task-1266549" [ 600.852868] env[62558]: _type = "Task" [ 600.852868] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.861366] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266549, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.239977] env[62558]: DEBUG nova.compute.utils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.244858] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 601.244945] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 601.282734] env[62558]: DEBUG nova.network.neutron [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.312410] env[62558]: DEBUG nova.policy [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36729b6b4df2408f98374f00a6b34e96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ab6d633c1cd4abaa63566ab5a0b2429', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.317912] env[62558]: DEBUG nova.network.neutron [-] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.353811] env[62558]: INFO nova.scheduler.client.report [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Deleted allocations for instance f52921c6-abbc-4a27-a451-e1cb4f5dfc48 [ 601.376879] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266549, 'name': Rename_Task, 'duration_secs': 0.147188} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.376879] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 601.376986] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b555b051-4f37-4ca6-abcb-b86b5a4ca3c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.385790] env[62558]: DEBUG nova.network.neutron [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.388965] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 601.388965] env[62558]: value = "task-1266550" [ 601.388965] env[62558]: _type = "Task" [ 601.388965] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.398165] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.681770] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa7e84b-a130-4f82-b292-a3c87152cd8a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.689287] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006d44ab-b29d-43ac-9634-3bff70a67528 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.722817] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d60257-077a-4951-ae94-73b0dcc3459f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.730193] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1dff8a-a837-43c2-a200-bb3fa3f97293 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.744334] env[62558]: DEBUG nova.compute.provider_tree [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.748558] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.823282] env[62558]: INFO nova.compute.manager [-] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Took 1.02 seconds to deallocate network for instance. [ 601.827668] env[62558]: DEBUG nova.compute.claims [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 601.829018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.870020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e3be2db-83e3-4e5b-8b47-9d85dc05412e tempest-FloatingIPsAssociationNegativeTestJSON-299379640 tempest-FloatingIPsAssociationNegativeTestJSON-299379640-project-member] Lock "f52921c6-abbc-4a27-a451-e1cb4f5dfc48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.312s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.878104] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Successfully created port: 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.890293] env[62558]: DEBUG oslo_concurrency.lockutils [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] Releasing lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.890813] env[62558]: DEBUG nova.compute.manager [req-24da4a3b-f1ae-4db3-b01c-4c527e31d52e req-3416d33d-711b-454a-8b45-963a88e7240b service nova] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Received event network-vif-deleted-053e1b3b-2d48-4717-8475-9241c88ba08e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.900030] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266550, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.251763] env[62558]: DEBUG nova.scheduler.client.report [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.372536] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.401560] env[62558]: DEBUG oslo_vmware.api [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266550, 'name': PowerOnVM_Task, 'duration_secs': 1.001095} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.401560] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 602.401560] env[62558]: DEBUG nova.compute.manager [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 602.402483] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da60b6c-841b-4001-b047-d5bc55a3abc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.765026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.765026] env[62558]: ERROR nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Traceback (most recent call last): [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.driver.spawn(context, instance, image_meta, [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.765026] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] vm_ref = self.build_virtual_machine(instance, [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] for vif in network_info: [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self._sync_wrapper(fn, *args, **kwargs) [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.wait() [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self[:] = self._gt.wait() [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self._exit_event.wait() [ 602.765748] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] result = hub.switch() [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return self.greenlet.switch() [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] result = function(*args, **kwargs) [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] return func(*args, **kwargs) [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise e [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] nwinfo = self.network_api.allocate_for_instance( [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.766130] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] created_port_ids = self._update_ports_for_instance( [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] with excutils.save_and_reraise_exception(): [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] self.force_reraise() [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise self.value [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] updated_port = self._update_port( [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] _ensure_no_port_binding_failure(port) [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.766583] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] raise exception.PortBindingFailed(port_id=port['id']) [ 602.766927] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] nova.exception.PortBindingFailed: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. [ 602.766927] env[62558]: ERROR nova.compute.manager [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] [ 602.766927] env[62558]: DEBUG nova.compute.utils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.769636] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.774907] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Build of instance 5fc9a508-323d-4f1c-ae46-d15f340f6b73 was re-scheduled: Binding failed for port 2a94ba19-7663-49a9-90cf-06b74bb6525b, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.775438] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.775793] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquiring lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.776046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Acquired lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.776520] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.780041] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.172s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.780041] env[62558]: DEBUG nova.objects.instance [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 602.812413] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.812814] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.813524] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.813524] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.813524] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.813987] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.814276] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.816031] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.816031] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.816031] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.816031] env[62558]: DEBUG nova.virt.hardware [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.816839] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697251aa-145a-4ffd-b945-7f672569afb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.826361] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0e6466-605c-47b9-8524-e33c00d5a248 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.899923] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.925276] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.053297] env[62558]: DEBUG nova.compute.manager [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Received event network-changed-7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.053297] env[62558]: DEBUG nova.compute.manager [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Refreshing instance network info cache due to event network-changed-7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.053297] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] Acquiring lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.053297] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] Acquired lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.053297] env[62558]: DEBUG nova.network.neutron [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Refreshing network info cache for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.221601] env[62558]: ERROR nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 603.221601] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.221601] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.221601] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.221601] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.221601] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.221601] env[62558]: ERROR nova.compute.manager raise self.value [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.221601] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.221601] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.221601] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.222052] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.222052] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.222052] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 603.222052] env[62558]: ERROR nova.compute.manager [ 603.222052] env[62558]: Traceback (most recent call last): [ 603.222052] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.222052] env[62558]: listener.cb(fileno) [ 603.222052] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.222052] env[62558]: result = function(*args, **kwargs) [ 603.222052] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.222052] env[62558]: return func(*args, **kwargs) [ 603.222052] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.222052] env[62558]: raise e [ 603.222052] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.222052] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 603.222052] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.222052] env[62558]: created_port_ids = self._update_ports_for_instance( [ 603.222052] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.222052] env[62558]: with excutils.save_and_reraise_exception(): [ 603.222484] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.222484] env[62558]: self.force_reraise() [ 603.222484] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.222484] env[62558]: raise self.value [ 603.222484] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.222484] env[62558]: updated_port = self._update_port( [ 603.222484] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.222484] env[62558]: _ensure_no_port_binding_failure(port) [ 603.222484] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.222484] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.222484] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 603.222484] env[62558]: Removing descriptor: 22 [ 603.223669] env[62558]: ERROR nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Traceback (most recent call last): [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] yield resources [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.driver.spawn(context, instance, image_meta, [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] vm_ref = self.build_virtual_machine(instance, [ 603.223669] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] for vif in network_info: [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self._sync_wrapper(fn, *args, **kwargs) [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.wait() [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self[:] = self._gt.wait() [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self._exit_event.wait() [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.224030] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] result = hub.switch() [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self.greenlet.switch() [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] result = function(*args, **kwargs) [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return func(*args, **kwargs) [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise e [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] nwinfo = self.network_api.allocate_for_instance( [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] created_port_ids = self._update_ports_for_instance( [ 603.224362] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] with excutils.save_and_reraise_exception(): [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.force_reraise() [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise self.value [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] updated_port = self._update_port( [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] _ensure_no_port_binding_failure(port) [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise exception.PortBindingFailed(port_id=port['id']) [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 603.224751] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] [ 603.225072] env[62558]: INFO nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Terminating instance [ 603.226362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.301390] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.419797] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.501377] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "f4f7e2c9-229c-4547-b674-5a5219a28151" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.501932] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.502303] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "f4f7e2c9-229c-4547-b674-5a5219a28151-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.502839] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.503209] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.505952] env[62558]: INFO nova.compute.manager [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Terminating instance [ 604.109434] env[62558]: DEBUG oslo_concurrency.lockutils [None req-db8ee7cc-0274-4cf8-9ed6-8239463abbdd tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.331s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.112149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Releasing lock "refresh_cache-5fc9a508-323d-4f1c-ae46-d15f340f6b73" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.112149] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 604.112149] env[62558]: DEBUG nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.112149] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.112740] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.112893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquired lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.113059] env[62558]: DEBUG nova.network.neutron [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.115896] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.908s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.131073] env[62558]: DEBUG nova.network.neutron [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.137179] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.228027] env[62558]: DEBUG nova.network.neutron [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.635517] env[62558]: DEBUG nova.network.neutron [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.638828] env[62558]: DEBUG nova.network.neutron [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.693670] env[62558]: DEBUG nova.network.neutron [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.732014] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c454c61-2c3e-4910-a2d6-113087077eee req-0e3b8ba0-090e-4996-8b42-421330c75910 service nova] Releasing lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.732014] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquired lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.732014] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.066969] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259ffba5-98af-46bb-aa3d-45ea192814b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.075222] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9867452d-2a01-47c4-8d02-5c0011cecc5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.107975] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4a1aec-e5e1-450b-9a20-ed014170398d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.112510] env[62558]: DEBUG nova.compute.manager [req-1847772f-26a8-4fb4-b8fe-41058a5eb303 req-edf4255a-5ff7-4a31-a28b-1027b018a9e3 service nova] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Received event network-vif-deleted-7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.118599] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23968072-0284-4e80-83d4-382ede04861f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.135804] env[62558]: DEBUG nova.compute.provider_tree [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.143492] env[62558]: INFO nova.compute.manager [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] [instance: 5fc9a508-323d-4f1c-ae46-d15f340f6b73] Took 1.03 seconds to deallocate network for instance. [ 605.195946] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Releasing lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.196402] env[62558]: DEBUG nova.compute.manager [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.196639] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.197792] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adc9a81-00b8-46b1-88f7-77302776b494 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.207324] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 605.207592] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8008725-1468-42d4-9bbc-a5f93bb9d54b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.215099] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 605.215099] env[62558]: value = "task-1266551" [ 605.215099] env[62558]: _type = "Task" [ 605.215099] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.222823] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.251100] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.333297] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.639031] env[62558]: DEBUG nova.scheduler.client.report [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.725823] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266551, 'name': PowerOffVM_Task, 'duration_secs': 0.186127} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.726104] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 605.726273] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 605.726512] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-140489ef-02bb-49ba-ad83-873c8dc88048 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.752407] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 605.752529] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 605.752699] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Deleting the datastore file [datastore2] f4f7e2c9-229c-4547-b674-5a5219a28151 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 605.753021] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d22e051-2950-4163-bc7d-7dd65ca71955 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.759955] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for the task: (returnval){ [ 605.759955] env[62558]: value = "task-1266553" [ 605.759955] env[62558]: _type = "Task" [ 605.759955] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.768080] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.837252] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Releasing lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.838295] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.838698] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.839232] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02293750-08c8-45fd-a008-9d8c22e86f2b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.851269] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6978ab09-002f-49d5-b0ec-5af454a437dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.876534] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76 could not be found. [ 605.876785] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.876964] env[62558]: INFO nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Took 0.04 seconds to destroy the instance on the hypervisor. [ 605.877217] env[62558]: DEBUG oslo.service.loopingcall [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.877444] env[62558]: DEBUG nova.compute.manager [-] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.877503] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.897495] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.113355] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "151768fd-c2c8-491f-93b5-bbc53bd69e2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.113654] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "151768fd-c2c8-491f-93b5-bbc53bd69e2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.146942] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.147584] env[62558]: ERROR nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Traceback (most recent call last): [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.driver.spawn(context, instance, image_meta, [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] vm_ref = self.build_virtual_machine(instance, [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.147584] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] for vif in network_info: [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self._sync_wrapper(fn, *args, **kwargs) [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.wait() [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self[:] = self._gt.wait() [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self._exit_event.wait() [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] result = hub.switch() [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.147922] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return self.greenlet.switch() [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] result = function(*args, **kwargs) [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] return func(*args, **kwargs) [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise e [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] nwinfo = self.network_api.allocate_for_instance( [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] created_port_ids = self._update_ports_for_instance( [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] with excutils.save_and_reraise_exception(): [ 606.148288] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] self.force_reraise() [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise self.value [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] updated_port = self._update_port( [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] _ensure_no_port_binding_failure(port) [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] raise exception.PortBindingFailed(port_id=port['id']) [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] nova.exception.PortBindingFailed: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. [ 606.148639] env[62558]: ERROR nova.compute.manager [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] [ 606.148939] env[62558]: DEBUG nova.compute.utils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.150574] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.826s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.153794] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Build of instance 5845d273-b18a-447a-ad2d-a011bfcc2423 was re-scheduled: Binding failed for port c83e936a-60b1-40bd-ba75-3d7c01abc582, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 606.154586] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 606.154813] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquiring lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.155088] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Acquired lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.155268] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.183290] env[62558]: INFO nova.scheduler.client.report [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Deleted allocations for instance 5fc9a508-323d-4f1c-ae46-d15f340f6b73 [ 606.271294] env[62558]: DEBUG oslo_vmware.api [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Task: {'id': task-1266553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100401} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.271566] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 606.271750] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 606.271919] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.272104] env[62558]: INFO nova.compute.manager [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Took 1.08 seconds to destroy the instance on the hypervisor. [ 606.272372] env[62558]: DEBUG oslo.service.loopingcall [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.272527] env[62558]: DEBUG nova.compute.manager [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.272619] env[62558]: DEBUG nova.network.neutron [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.288446] env[62558]: DEBUG nova.network.neutron [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.400031] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.441188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "f3128303-9b2c-4f14-9ea0-093f55ecb063" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.441414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "f3128303-9b2c-4f14-9ea0-093f55ecb063" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.687317] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.693919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3df11a8d-e83e-4258-b9aa-ebdd4909d7c0 tempest-InstanceActionsV221TestJSON-559300460 tempest-InstanceActionsV221TestJSON-559300460-project-member] Lock "5fc9a508-323d-4f1c-ae46-d15f340f6b73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.170s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.790065] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.791255] env[62558]: DEBUG nova.network.neutron [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.902357] env[62558]: INFO nova.compute.manager [-] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Took 1.02 seconds to deallocate network for instance. [ 606.904857] env[62558]: DEBUG nova.compute.claims [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.905041] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.124258] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f8b4f9-c465-4bea-91f1-617d193ea356 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.138959] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bc6f2f-c728-4ea7-95c9-46b6979753ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.173461] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e091a34c-71b6-4684-ba4e-9b793082a453 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.184984] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b75b83-74cc-4e23-8e72-6a246d38d0e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.198820] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.201746] env[62558]: DEBUG nova.compute.provider_tree [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.293487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Releasing lock "refresh_cache-5845d273-b18a-447a-ad2d-a011bfcc2423" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.294192] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 607.294408] env[62558]: DEBUG nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.294608] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.299997] env[62558]: INFO nova.compute.manager [-] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Took 1.03 seconds to deallocate network for instance. [ 607.314784] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.565547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquiring lock "3e035ee1-a63a-46a6-84ca-4980e1185e03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.565980] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "3e035ee1-a63a-46a6-84ca-4980e1185e03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.706960] env[62558]: DEBUG nova.scheduler.client.report [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.729349] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.808133] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.817914] env[62558]: DEBUG nova.network.neutron [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.214256] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.215397] env[62558]: ERROR nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Traceback (most recent call last): [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.driver.spawn(context, instance, image_meta, [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] vm_ref = self.build_virtual_machine(instance, [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.215397] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] for vif in network_info: [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self._sync_wrapper(fn, *args, **kwargs) [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.wait() [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self[:] = self._gt.wait() [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self._exit_event.wait() [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] result = hub.switch() [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.215976] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return self.greenlet.switch() [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] result = function(*args, **kwargs) [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] return func(*args, **kwargs) [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise e [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] nwinfo = self.network_api.allocate_for_instance( [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] created_port_ids = self._update_ports_for_instance( [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] with excutils.save_and_reraise_exception(): [ 608.216733] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] self.force_reraise() [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise self.value [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] updated_port = self._update_port( [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] _ensure_no_port_binding_failure(port) [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] raise exception.PortBindingFailed(port_id=port['id']) [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] nova.exception.PortBindingFailed: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. [ 608.217443] env[62558]: ERROR nova.compute.manager [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] [ 608.218096] env[62558]: DEBUG nova.compute.utils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.218096] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.402s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.218282] env[62558]: DEBUG nova.objects.instance [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lazy-loading 'resources' on Instance uuid 9316facd-0c95-4720-b0d3-c0a39df81619 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 608.220361] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Build of instance 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f was re-scheduled: Binding failed for port 2a1bff7c-65d5-4198-99c0-4566c13d38be, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.220924] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.220924] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.221084] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquired lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.221185] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.320561] env[62558]: INFO nova.compute.manager [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] [instance: 5845d273-b18a-447a-ad2d-a011bfcc2423] Took 1.03 seconds to deallocate network for instance. [ 608.741078] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.823590] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.196504] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b029ce11-e742-4947-947a-ca7483d736dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.206185] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbaee24-1be2-4118-b7af-0add333c134c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.244712] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53675c1d-2b96-4e64-a3e1-686680feab34 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.254606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0305b6-1c28-4d75-8ea2-eed93858160a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.270452] env[62558]: DEBUG nova.compute.provider_tree [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.330232] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Releasing lock "refresh_cache-37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.330429] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 609.330619] env[62558]: DEBUG nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.330785] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.356686] env[62558]: INFO nova.scheduler.client.report [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Deleted allocations for instance 5845d273-b18a-447a-ad2d-a011bfcc2423 [ 609.363029] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.774389] env[62558]: DEBUG nova.scheduler.client.report [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.865799] env[62558]: DEBUG nova.network.neutron [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.868400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e698fa2-91db-4641-910b-07e4b1d868d9 tempest-ServerExternalEventsTest-1021826053 tempest-ServerExternalEventsTest-1021826053-project-member] Lock "5845d273-b18a-447a-ad2d-a011bfcc2423" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.695s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.279784] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.062s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.282861] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.517s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.304888] env[62558]: INFO nova.scheduler.client.report [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Deleted allocations for instance 9316facd-0c95-4720-b0d3-c0a39df81619 [ 610.369366] env[62558]: INFO nova.compute.manager [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f] Took 1.04 seconds to deallocate network for instance. [ 610.373573] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.812715] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74d29900-dead-4722-8aeb-82acdce311f4 tempest-ServerShowV257Test-665404909 tempest-ServerShowV257Test-665404909-project-member] Lock "9316facd-0c95-4720-b0d3-c0a39df81619" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.809s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.903098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.272532] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2d1189-7860-4f57-953d-fe39d439e8ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.280679] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2597fd4f-ebdc-4785-8ffc-caa825455f7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.310248] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db02fab-d6e0-4545-ab54-3b9673e55d86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.317966] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99dfd695-8a98-4b92-a5b3-ba5c1cb538a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.332486] env[62558]: DEBUG nova.compute.provider_tree [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.403692] env[62558]: INFO nova.scheduler.client.report [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Deleted allocations for instance 37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f [ 611.839501] env[62558]: DEBUG nova.scheduler.client.report [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.914587] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2e53a5a3-1da0-4edf-9bf1-dd2abd23d5e0 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "37cbd58c-b11a-4a62-95b2-5f4c5cf7ad1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.243s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.349831] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.350496] env[62558]: ERROR nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Traceback (most recent call last): [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.driver.spawn(context, instance, image_meta, [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] vm_ref = self.build_virtual_machine(instance, [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.350496] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] for vif in network_info: [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self._sync_wrapper(fn, *args, **kwargs) [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.wait() [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self[:] = self._gt.wait() [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self._exit_event.wait() [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] result = hub.switch() [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.350852] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return self.greenlet.switch() [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] result = function(*args, **kwargs) [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] return func(*args, **kwargs) [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise e [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] nwinfo = self.network_api.allocate_for_instance( [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] created_port_ids = self._update_ports_for_instance( [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] with excutils.save_and_reraise_exception(): [ 612.351229] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] self.force_reraise() [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise self.value [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] updated_port = self._update_port( [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] _ensure_no_port_binding_failure(port) [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] raise exception.PortBindingFailed(port_id=port['id']) [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] nova.exception.PortBindingFailed: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. [ 612.351732] env[62558]: ERROR nova.compute.manager [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] [ 612.352036] env[62558]: DEBUG nova.compute.utils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.352498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.869s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.353916] env[62558]: INFO nova.compute.claims [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.356550] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Build of instance 813e287c-9d7f-4e9c-9216-986f631b6eb6 was re-scheduled: Binding failed for port 1d437e03-f755-47ff-9d65-0951fcbd5722, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.357034] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.357335] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquiring lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.357431] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Acquired lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.357553] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.419071] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.885698] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.944919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.008674] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.314247] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquiring lock "d526aa96-4007-45f8-8eee-9030bdb9f402" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.314482] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "d526aa96-4007-45f8-8eee-9030bdb9f402" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.514773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Releasing lock "refresh_cache-813e287c-9d7f-4e9c-9216-986f631b6eb6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.515058] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.515271] env[62558]: DEBUG nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.515461] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.543758] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.842250] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1181b02c-f827-4b20-b199-b2093e480298 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.850613] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e07555-79c5-4c59-aaf9-5c0f9c983684 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.884358] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee911e1a-04b2-4e2a-9d94-3d0a00a9b65f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.892274] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bbd04c-0f02-4b66-b440-a989c7b9a565 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.906353] env[62558]: DEBUG nova.compute.provider_tree [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.049515] env[62558]: DEBUG nova.network.neutron [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.412174] env[62558]: DEBUG nova.scheduler.client.report [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.552489] env[62558]: INFO nova.compute.manager [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] [instance: 813e287c-9d7f-4e9c-9216-986f631b6eb6] Took 1.04 seconds to deallocate network for instance. [ 614.914413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.915133] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.918270] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.819s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.394441] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.394441] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.425024] env[62558]: DEBUG nova.compute.utils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.425024] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.425024] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.554792] env[62558]: DEBUG nova.policy [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1ada4fb2d454435a1aa8f83ab9d35d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddad32020e004244a0f15141ba838c2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.598128] env[62558]: INFO nova.scheduler.client.report [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Deleted allocations for instance 813e287c-9d7f-4e9c-9216-986f631b6eb6 [ 615.902649] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.902649] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 615.902649] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 615.930502] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.955370] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6049a2-72e1-416b-b7ee-84184899a0fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.963456] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc867a00-d1f2-44e5-bcf0-ffeb173fd4bf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.999537] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e46ed26-54a6-4998-90e9-4ad9872eee77 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.009394] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a716c026-06f0-47ee-ac0f-d6a26b2d323f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.025712] env[62558]: DEBUG nova.compute.provider_tree [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.111112] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0802d429-3e13-44df-b39c-20df94c173d6 tempest-AttachInterfacesUnderV243Test-1300824439 tempest-AttachInterfacesUnderV243Test-1300824439-project-member] Lock "813e287c-9d7f-4e9c-9216-986f631b6eb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.674s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.205853] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Successfully created port: 43fbab1a-a239-4a31-abca-4451749b924d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.405168] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.405349] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.405483] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.405609] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.438963] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.438963] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.438963] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 616.438963] env[62558]: DEBUG nova.objects.instance [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lazy-loading 'info_cache' on Instance uuid f4f7e2c9-229c-4547-b674-5a5219a28151 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.529802] env[62558]: DEBUG nova.scheduler.client.report [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.614444] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.942302] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.977023] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.977023] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.977023] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.977309] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.977309] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.977309] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.977309] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.977309] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.977440] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.977440] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.977440] env[62558]: DEBUG nova.virt.hardware [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.977729] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b20ac7-3d61-43ff-97f5-fca28696d24a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.986529] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d84c297-da2d-41f9-a3e0-f1f235421f32 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.036178] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.118s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.036938] env[62558]: ERROR nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Traceback (most recent call last): [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.driver.spawn(context, instance, image_meta, [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] vm_ref = self.build_virtual_machine(instance, [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.036938] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] for vif in network_info: [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self._sync_wrapper(fn, *args, **kwargs) [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.wait() [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self[:] = self._gt.wait() [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self._exit_event.wait() [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] result = hub.switch() [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.037336] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return self.greenlet.switch() [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] result = function(*args, **kwargs) [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] return func(*args, **kwargs) [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise e [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] nwinfo = self.network_api.allocate_for_instance( [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] created_port_ids = self._update_ports_for_instance( [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] with excutils.save_and_reraise_exception(): [ 617.038799] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] self.force_reraise() [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise self.value [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] updated_port = self._update_port( [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] _ensure_no_port_binding_failure(port) [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] raise exception.PortBindingFailed(port_id=port['id']) [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] nova.exception.PortBindingFailed: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. [ 617.039106] env[62558]: ERROR nova.compute.manager [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] [ 617.039382] env[62558]: DEBUG nova.compute.utils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.039382] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.263s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.046900] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Build of instance 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e was re-scheduled: Binding failed for port 5d42bc9c-0db8-4264-a494-384dcc6f36dc, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.046900] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.046900] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.046900] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquired lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.047143] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.138143] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.475057] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.483879] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "d92dca09-0a42-4103-ac92-25b6b0fa3c32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.484128] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "d92dca09-0a42-4103-ac92-25b6b0fa3c32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.579574] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.667283] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.790667] env[62558]: ERROR nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 617.790667] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.790667] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.790667] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.790667] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.790667] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.790667] env[62558]: ERROR nova.compute.manager raise self.value [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.790667] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.790667] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.790667] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.791149] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.791149] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.791149] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 617.791149] env[62558]: ERROR nova.compute.manager [ 617.791149] env[62558]: Traceback (most recent call last): [ 617.791149] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.791149] env[62558]: listener.cb(fileno) [ 617.791149] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.791149] env[62558]: result = function(*args, **kwargs) [ 617.791149] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.791149] env[62558]: return func(*args, **kwargs) [ 617.791149] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.791149] env[62558]: raise e [ 617.791149] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.791149] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 617.791149] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.791149] env[62558]: created_port_ids = self._update_ports_for_instance( [ 617.791149] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.791149] env[62558]: with excutils.save_and_reraise_exception(): [ 617.791149] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.791149] env[62558]: self.force_reraise() [ 617.791149] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.791149] env[62558]: raise self.value [ 617.791149] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.791149] env[62558]: updated_port = self._update_port( [ 617.791149] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.791149] env[62558]: _ensure_no_port_binding_failure(port) [ 617.791149] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.791149] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.791823] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 617.791823] env[62558]: Removing descriptor: 22 [ 617.791823] env[62558]: ERROR nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Traceback (most recent call last): [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] yield resources [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.driver.spawn(context, instance, image_meta, [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.791823] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] vm_ref = self.build_virtual_machine(instance, [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] for vif in network_info: [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self._sync_wrapper(fn, *args, **kwargs) [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.wait() [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self[:] = self._gt.wait() [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self._exit_event.wait() [ 617.792241] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] result = hub.switch() [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self.greenlet.switch() [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] result = function(*args, **kwargs) [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return func(*args, **kwargs) [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise e [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] nwinfo = self.network_api.allocate_for_instance( [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.792537] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] created_port_ids = self._update_ports_for_instance( [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] with excutils.save_and_reraise_exception(): [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.force_reraise() [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise self.value [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] updated_port = self._update_port( [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] _ensure_no_port_binding_failure(port) [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.792845] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise exception.PortBindingFailed(port_id=port['id']) [ 617.793129] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 617.793129] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] [ 617.793129] env[62558]: INFO nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Terminating instance [ 617.796764] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquiring lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.797942] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquired lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.797942] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.825339] env[62558]: DEBUG nova.compute.manager [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Received event network-changed-43fbab1a-a239-4a31-abca-4451749b924d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.825439] env[62558]: DEBUG nova.compute.manager [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Refreshing instance network info cache due to event network-changed-43fbab1a-a239-4a31-abca-4451749b924d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.825661] env[62558]: DEBUG oslo_concurrency.lockutils [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] Acquiring lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.961263] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5116e9-b36b-492e-b72b-5162d47853c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.969327] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2642939d-4074-4b0a-9d23-f85841d39b01 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.001367] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88410915-9e99-43be-a8f7-e73297f5670d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.009843] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80dad836-49c6-4c92-aedd-d59ac53f3dd1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.023835] env[62558]: DEBUG nova.compute.provider_tree [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.128438] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.169608] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Releasing lock "refresh_cache-14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.169847] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.170042] env[62558]: DEBUG nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.170215] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.199823] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.323135] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.403104] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.527423] env[62558]: DEBUG nova.scheduler.client.report [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.630067] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-f4f7e2c9-229c-4547-b674-5a5219a28151" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.630329] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 618.630494] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.630663] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.630819] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.630982] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.631150] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.631391] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.631472] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 618.631582] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.704042] env[62558]: DEBUG nova.network.neutron [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.906639] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Releasing lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.907319] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.907704] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.911022] env[62558]: DEBUG oslo_concurrency.lockutils [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] Acquired lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.911022] env[62558]: DEBUG nova.network.neutron [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Refreshing network info cache for port 43fbab1a-a239-4a31-abca-4451749b924d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.911022] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7e262b9-d936-41b5-b4f7-26cb458a3fc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.924849] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02c90e7-d198-48d3-bb49-64af3d399882 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.949955] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e1410c0-8aff-429c-8bbb-46d9bc01de32 could not be found. [ 618.949955] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.949955] env[62558]: INFO nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.950278] env[62558]: DEBUG oslo.service.loopingcall [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.950343] env[62558]: DEBUG nova.compute.manager [-] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.950435] env[62558]: DEBUG nova.network.neutron [-] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.965416] env[62558]: DEBUG nova.network.neutron [-] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.033389] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.034075] env[62558]: ERROR nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Traceback (most recent call last): [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.driver.spawn(context, instance, image_meta, [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] vm_ref = self.build_virtual_machine(instance, [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.034075] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] for vif in network_info: [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self._sync_wrapper(fn, *args, **kwargs) [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.wait() [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self[:] = self._gt.wait() [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self._exit_event.wait() [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] result = hub.switch() [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.034424] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return self.greenlet.switch() [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] result = function(*args, **kwargs) [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] return func(*args, **kwargs) [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise e [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] nwinfo = self.network_api.allocate_for_instance( [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] created_port_ids = self._update_ports_for_instance( [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] with excutils.save_and_reraise_exception(): [ 619.034709] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] self.force_reraise() [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise self.value [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] updated_port = self._update_port( [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] _ensure_no_port_binding_failure(port) [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] raise exception.PortBindingFailed(port_id=port['id']) [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] nova.exception.PortBindingFailed: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. [ 619.034999] env[62558]: ERROR nova.compute.manager [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] [ 619.035273] env[62558]: DEBUG nova.compute.utils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.036061] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.208s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.040851] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Build of instance 5a1f074f-8e43-42f9-864d-976cdaa4639c was re-scheduled: Binding failed for port c85a916b-8b11-45f3-ac31-b803326f3461, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.041660] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.041660] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquiring lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.041660] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Acquired lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.041660] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.134784] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.207070] env[62558]: INFO nova.compute.manager [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e] Took 1.04 seconds to deallocate network for instance. [ 619.437109] env[62558]: DEBUG nova.network.neutron [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.469470] env[62558]: DEBUG nova.network.neutron [-] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.514243] env[62558]: DEBUG nova.network.neutron [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.573963] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.701511] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.948310] env[62558]: DEBUG nova.compute.manager [req-c7c66ebb-4e20-4fa1-8dc8-79de3ac7225f req-a1afa3e7-6a85-427c-be68-5baf2639d9cb service nova] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Received event network-vif-deleted-43fbab1a-a239-4a31-abca-4451749b924d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.972473] env[62558]: INFO nova.compute.manager [-] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Took 1.02 seconds to deallocate network for instance. [ 619.976571] env[62558]: DEBUG nova.compute.claims [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.976761] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.017028] env[62558]: DEBUG oslo_concurrency.lockutils [req-77f6f2ab-1ea0-4bdd-84f7-ecb7de367430 req-3487364f-73df-4e1b-83e4-1196c2f36048 service nova] Releasing lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.035277] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba4c458-f6d5-4de9-a6f0-8bbd3c8066b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.043293] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9cb1b5-4a3d-4bb1-b0f8-13e451e65964 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.075569] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c5b18d-4599-48cd-b2f5-e475306b5d1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.082935] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a01994-c5e6-4ae7-9e22-26483de901b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.097426] env[62558]: DEBUG nova.compute.provider_tree [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.205312] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Releasing lock "refresh_cache-5a1f074f-8e43-42f9-864d-976cdaa4639c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.205614] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.205893] env[62558]: DEBUG nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.206157] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.225694] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.245287] env[62558]: INFO nova.scheduler.client.report [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Deleted allocations for instance 14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e [ 620.591763] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.592063] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.602650] env[62558]: DEBUG nova.scheduler.client.report [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.729125] env[62558]: DEBUG nova.network.neutron [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.754070] env[62558]: DEBUG oslo_concurrency.lockutils [None req-057464a4-a7eb-4183-9123-48dfa201a5f9 tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "14c27acb-0c19-47f9-b3fd-4f9e3e83bb5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.258s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.107769] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.109032] env[62558]: ERROR nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Traceback (most recent call last): [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.driver.spawn(context, instance, image_meta, [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] vm_ref = self.build_virtual_machine(instance, [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.109032] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] for vif in network_info: [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self._sync_wrapper(fn, *args, **kwargs) [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.wait() [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self[:] = self._gt.wait() [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self._exit_event.wait() [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] result = hub.switch() [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.109337] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return self.greenlet.switch() [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] result = function(*args, **kwargs) [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] return func(*args, **kwargs) [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise e [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] nwinfo = self.network_api.allocate_for_instance( [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] created_port_ids = self._update_ports_for_instance( [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] with excutils.save_and_reraise_exception(): [ 621.109628] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] self.force_reraise() [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise self.value [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] updated_port = self._update_port( [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] _ensure_no_port_binding_failure(port) [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] raise exception.PortBindingFailed(port_id=port['id']) [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] nova.exception.PortBindingFailed: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. [ 621.109936] env[62558]: ERROR nova.compute.manager [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] [ 621.110409] env[62558]: DEBUG nova.compute.utils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.111731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.213s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.113500] env[62558]: INFO nova.compute.claims [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.116576] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Build of instance 1513843d-0a96-4b56-86bb-3c39ab4b3692 was re-scheduled: Binding failed for port 053e1b3b-2d48-4717-8475-9241c88ba08e, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.117200] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.117370] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquiring lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.117524] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Acquired lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.117679] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.233306] env[62558]: INFO nova.compute.manager [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] [instance: 5a1f074f-8e43-42f9-864d-976cdaa4639c] Took 1.03 seconds to deallocate network for instance. [ 621.258188] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.646426] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.749408] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.776580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "d65fbfe3-b1e9-48d9-88ba-209fdc264283" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.776986] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "d65fbfe3-b1e9-48d9-88ba-209fdc264283" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.785423] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.195198] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.195464] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.251484] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Releasing lock "refresh_cache-1513843d-0a96-4b56-86bb-3c39ab4b3692" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.251659] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.251835] env[62558]: DEBUG nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.251993] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.265671] env[62558]: INFO nova.scheduler.client.report [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Deleted allocations for instance 5a1f074f-8e43-42f9-864d-976cdaa4639c [ 622.275189] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.569299] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d9e9d9-3f65-4b28-807e-f031cb060ad0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.578217] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2eba80-811c-4197-b098-6e3de2585715 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.613831] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3042e86-e0f9-466c-a381-0b25ecfcbbcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.619625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.619836] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.626500] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8271d8-fea5-47c7-8c7b-2032b144a687 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.640410] env[62558]: DEBUG nova.compute.provider_tree [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.777737] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f1380d00-3367-40ce-bbb1-399dc23dfd5d tempest-ServerAddressesNegativeTestJSON-323139077 tempest-ServerAddressesNegativeTestJSON-323139077-project-member] Lock "5a1f074f-8e43-42f9-864d-976cdaa4639c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.118s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.779021] env[62558]: DEBUG nova.network.neutron [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.041066] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "b0038711-5d12-4909-b331-72acb5ed0d24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.041066] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.146186] env[62558]: DEBUG nova.scheduler.client.report [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.282770] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.284898] env[62558]: INFO nova.compute.manager [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] [instance: 1513843d-0a96-4b56-86bb-3c39ab4b3692] Took 1.03 seconds to deallocate network for instance. [ 623.652908] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.653484] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.656131] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.731s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.656380] env[62558]: DEBUG nova.objects.instance [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 623.816795] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.160837] env[62558]: DEBUG nova.compute.utils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.165818] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.165985] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.230927] env[62558]: DEBUG nova.policy [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '969e0fdf86264a16ab12a43287433f0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8166a9c95d364dc58126f7eaef3c6761', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.328668] env[62558]: INFO nova.scheduler.client.report [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Deleted allocations for instance 1513843d-0a96-4b56-86bb-3c39ab4b3692 [ 624.667053] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.671669] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b40a2ffc-b2fd-4efd-8a20-43e88e1f866f tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.672815] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.768s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.751492] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Successfully created port: 15f5b8c8-8fc2-4d2c-984e-9302eb332106 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.843131] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4269c4ed-6f93-4dc2-b429-665eb92a4067 tempest-ServersV294TestFqdnHostnames-963862238 tempest-ServersV294TestFqdnHostnames-963862238-project-member] Lock "1513843d-0a96-4b56-86bb-3c39ab4b3692" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.012s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.345848] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.599907] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4910d13-710c-424e-94d1-903cf8fa32ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.607929] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf265ca-320c-42bd-85e6-be297740620f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.637579] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc52c1f-0028-4264-9eb7-5b589e86f3d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.644668] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3732627-f162-4b8d-ae1e-65ccb3551423 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.658742] env[62558]: DEBUG nova.compute.provider_tree [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.681804] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.712019] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.712019] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.712019] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.712229] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.712518] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.712863] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.713225] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.713499] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.713773] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.716017] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.716017] env[62558]: DEBUG nova.virt.hardware [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.716017] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2321aff-0f40-48bd-ac29-feea7d312d9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.722753] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2351f9a9-c6be-4b55-9816-0bbf57c34683 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.871732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.875679] env[62558]: DEBUG nova.compute.manager [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Received event network-changed-15f5b8c8-8fc2-4d2c-984e-9302eb332106 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.875863] env[62558]: DEBUG nova.compute.manager [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Refreshing instance network info cache due to event network-changed-15f5b8c8-8fc2-4d2c-984e-9302eb332106. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.876084] env[62558]: DEBUG oslo_concurrency.lockutils [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] Acquiring lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.876224] env[62558]: DEBUG oslo_concurrency.lockutils [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] Acquired lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.876415] env[62558]: DEBUG nova.network.neutron [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Refreshing network info cache for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.959251] env[62558]: ERROR nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 625.959251] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.959251] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.959251] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.959251] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.959251] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.959251] env[62558]: ERROR nova.compute.manager raise self.value [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.959251] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.959251] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.959251] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.959734] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.959734] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.959734] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 625.959734] env[62558]: ERROR nova.compute.manager [ 625.959734] env[62558]: Traceback (most recent call last): [ 625.959734] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.959734] env[62558]: listener.cb(fileno) [ 625.959734] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.959734] env[62558]: result = function(*args, **kwargs) [ 625.959734] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.959734] env[62558]: return func(*args, **kwargs) [ 625.959734] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.959734] env[62558]: raise e [ 625.959734] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.959734] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 625.959734] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.959734] env[62558]: created_port_ids = self._update_ports_for_instance( [ 625.959734] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.959734] env[62558]: with excutils.save_and_reraise_exception(): [ 625.959734] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.959734] env[62558]: self.force_reraise() [ 625.959734] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.959734] env[62558]: raise self.value [ 625.959734] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.959734] env[62558]: updated_port = self._update_port( [ 625.959734] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.959734] env[62558]: _ensure_no_port_binding_failure(port) [ 625.959734] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.959734] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.960516] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 625.960516] env[62558]: Removing descriptor: 22 [ 625.960516] env[62558]: ERROR nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] Traceback (most recent call last): [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] yield resources [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.driver.spawn(context, instance, image_meta, [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.960516] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] vm_ref = self.build_virtual_machine(instance, [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] for vif in network_info: [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self._sync_wrapper(fn, *args, **kwargs) [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.wait() [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self[:] = self._gt.wait() [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self._exit_event.wait() [ 625.960851] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] result = hub.switch() [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self.greenlet.switch() [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] result = function(*args, **kwargs) [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return func(*args, **kwargs) [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise e [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] nwinfo = self.network_api.allocate_for_instance( [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.961197] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] created_port_ids = self._update_ports_for_instance( [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] with excutils.save_and_reraise_exception(): [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.force_reraise() [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise self.value [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] updated_port = self._update_port( [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] _ensure_no_port_binding_failure(port) [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.961550] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise exception.PortBindingFailed(port_id=port['id']) [ 625.961879] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 625.961879] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] [ 625.961879] env[62558]: INFO nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Terminating instance [ 625.962808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquiring lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.163848] env[62558]: DEBUG nova.scheduler.client.report [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.414923] env[62558]: DEBUG nova.network.neutron [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.522063] env[62558]: DEBUG nova.network.neutron [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.668420] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.669634] env[62558]: ERROR nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Traceback (most recent call last): [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.driver.spawn(context, instance, image_meta, [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] vm_ref = self.build_virtual_machine(instance, [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.669634] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] for vif in network_info: [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self._sync_wrapper(fn, *args, **kwargs) [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.wait() [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self[:] = self._gt.wait() [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self._exit_event.wait() [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] result = hub.switch() [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.669994] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return self.greenlet.switch() [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] result = function(*args, **kwargs) [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] return func(*args, **kwargs) [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise e [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] nwinfo = self.network_api.allocate_for_instance( [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] created_port_ids = self._update_ports_for_instance( [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] with excutils.save_and_reraise_exception(): [ 626.670374] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] self.force_reraise() [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise self.value [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] updated_port = self._update_port( [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] _ensure_no_port_binding_failure(port) [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] raise exception.PortBindingFailed(port_id=port['id']) [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] nova.exception.PortBindingFailed: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. [ 626.670782] env[62558]: ERROR nova.compute.manager [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] [ 626.671101] env[62558]: DEBUG nova.compute.utils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.671566] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.943s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.672972] env[62558]: INFO nova.compute.claims [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.675525] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Build of instance 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76 was re-scheduled: Binding failed for port 7dd86f2c-77c8-4b7b-8e81-3e692ad1cddd, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.675948] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.676181] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquiring lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.676324] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Acquired lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.676479] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.024405] env[62558]: DEBUG oslo_concurrency.lockutils [req-d7e72bfd-cf3e-46fb-b615-9b49766c7136 req-2d4997d9-fb03-4ccb-95f0-220ac362f19c service nova] Releasing lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.024834] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquired lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.025058] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.211611] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.278936] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.548251] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.631751] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.783442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Releasing lock "refresh_cache-4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.783685] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.783850] env[62558]: DEBUG nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.784444] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.809253] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.907902] env[62558]: DEBUG nova.compute.manager [req-a75805b9-5a9c-4928-8bfa-e5f4d029fe71 req-5526422d-67a3-4ee3-817e-225296c19811 service nova] [instance: 95348a45-f517-4e21-8969-57c978985984] Received event network-vif-deleted-15f5b8c8-8fc2-4d2c-984e-9302eb332106 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.090122] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e65373-d237-46de-bd30-c21026eaab5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.098198] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc0661d-288f-4192-9f4b-15fbc044c0b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.127993] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e94040-0e1e-41a3-afef-4cf9918bfccc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.134829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Releasing lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.135257] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.135485] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.136812] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb961258-eba2-4974-bef4-daa7a673db1c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.138626] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad87d43-28c1-45ff-908b-a8d62edf0ad5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.152034] env[62558]: DEBUG nova.compute.provider_tree [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.156775] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2fa1b2-04d2-4fe2-a1ac-1378e92ba28f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.179651] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95348a45-f517-4e21-8969-57c978985984 could not be found. [ 628.179884] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.180072] env[62558]: INFO nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Took 0.04 seconds to destroy the instance on the hypervisor. [ 628.180319] env[62558]: DEBUG oslo.service.loopingcall [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.180531] env[62558]: DEBUG nova.compute.manager [-] [instance: 95348a45-f517-4e21-8969-57c978985984] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.180621] env[62558]: DEBUG nova.network.neutron [-] [instance: 95348a45-f517-4e21-8969-57c978985984] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.194513] env[62558]: DEBUG nova.network.neutron [-] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.312537] env[62558]: DEBUG nova.network.neutron [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.657778] env[62558]: DEBUG nova.scheduler.client.report [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.697074] env[62558]: DEBUG nova.network.neutron [-] [instance: 95348a45-f517-4e21-8969-57c978985984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.814800] env[62558]: INFO nova.compute.manager [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] [instance: 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76] Took 1.03 seconds to deallocate network for instance. [ 629.162536] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.163108] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.166564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.359s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.166792] env[62558]: DEBUG nova.objects.instance [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lazy-loading 'resources' on Instance uuid f4f7e2c9-229c-4547-b674-5a5219a28151 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 629.199163] env[62558]: INFO nova.compute.manager [-] [instance: 95348a45-f517-4e21-8969-57c978985984] Took 1.02 seconds to deallocate network for instance. [ 629.201659] env[62558]: DEBUG nova.compute.claims [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.201837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.671016] env[62558]: DEBUG nova.compute.utils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.680972] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.680972] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 629.755045] env[62558]: DEBUG nova.policy [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c5503a5bf734483af68683a8c853a71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9d4de9f22ec414d90eb8c2ed9c42d18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.846078] env[62558]: INFO nova.scheduler.client.report [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Deleted allocations for instance 4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76 [ 630.122138] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23da98c-4474-4f11-a4ef-80bb3566fd95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.130071] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be5291f-554a-4268-9c31-bff3f77c680a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.172783] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5115621d-8610-4f89-b62f-23ff69105a22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.176194] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.182768] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afdc9adf-4dcc-405f-a0a2-a34af82da441 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.197231] env[62558]: DEBUG nova.compute.provider_tree [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.332509] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Successfully created port: dc2bdb70-ad34-41ce-b709-d7175df751ec {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.359641] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a426f4a4-49b3-4260-ad01-6ffeac4cce15 tempest-DeleteServersAdminTestJSON-2095886350 tempest-DeleteServersAdminTestJSON-2095886350-project-member] Lock "4e7e1c74-10a8-4746-b6a6-56dfcc0c2d76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.822s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.700952] env[62558]: DEBUG nova.scheduler.client.report [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.863138] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.187739] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.207494] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.214410] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.311s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.215974] env[62558]: INFO nova.compute.claims [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.227275] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.227544] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.227680] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.227856] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.228000] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.228239] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.228595] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.228681] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.228781] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.228948] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.229748] env[62558]: DEBUG nova.virt.hardware [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.232415] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb5c7bc-eb3d-4f78-a846-20ebc2bfba7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.241735] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd336a4-390e-4ec5-a784-a0c2e1e53438 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.248720] env[62558]: INFO nova.scheduler.client.report [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Deleted allocations for instance f4f7e2c9-229c-4547-b674-5a5219a28151 [ 631.389554] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.767143] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c3981956-0b4d-4be5-85f0-599b3b5bfb6d tempest-ServerShowV254Test-1724350 tempest-ServerShowV254Test-1724350-project-member] Lock "f4f7e2c9-229c-4547-b674-5a5219a28151" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.265s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.382739] env[62558]: DEBUG nova.compute.manager [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Received event network-changed-dc2bdb70-ad34-41ce-b709-d7175df751ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.383243] env[62558]: DEBUG nova.compute.manager [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Refreshing instance network info cache due to event network-changed-dc2bdb70-ad34-41ce-b709-d7175df751ec. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 632.383455] env[62558]: DEBUG oslo_concurrency.lockutils [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] Acquiring lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.383632] env[62558]: DEBUG oslo_concurrency.lockutils [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] Acquired lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.383880] env[62558]: DEBUG nova.network.neutron [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Refreshing network info cache for port dc2bdb70-ad34-41ce-b709-d7175df751ec {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.450077] env[62558]: ERROR nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 632.450077] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.450077] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.450077] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.450077] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.450077] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.450077] env[62558]: ERROR nova.compute.manager raise self.value [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.450077] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.450077] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.450077] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.450634] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.450634] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.450634] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 632.450634] env[62558]: ERROR nova.compute.manager [ 632.453759] env[62558]: Traceback (most recent call last): [ 632.453759] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.453759] env[62558]: listener.cb(fileno) [ 632.453759] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.453759] env[62558]: result = function(*args, **kwargs) [ 632.453759] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.453759] env[62558]: return func(*args, **kwargs) [ 632.453759] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.453759] env[62558]: raise e [ 632.453759] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.453759] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 632.453759] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.453759] env[62558]: created_port_ids = self._update_ports_for_instance( [ 632.453759] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.453759] env[62558]: with excutils.save_and_reraise_exception(): [ 632.453759] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.453759] env[62558]: self.force_reraise() [ 632.453759] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.453759] env[62558]: raise self.value [ 632.453759] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.453759] env[62558]: updated_port = self._update_port( [ 632.453759] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.453759] env[62558]: _ensure_no_port_binding_failure(port) [ 632.453759] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.453759] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.453759] env[62558]: nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 632.453759] env[62558]: Removing descriptor: 22 [ 632.454499] env[62558]: ERROR nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Traceback (most recent call last): [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] yield resources [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.driver.spawn(context, instance, image_meta, [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] vm_ref = self.build_virtual_machine(instance, [ 632.454499] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] for vif in network_info: [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self._sync_wrapper(fn, *args, **kwargs) [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.wait() [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self[:] = self._gt.wait() [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self._exit_event.wait() [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.454804] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] result = hub.switch() [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self.greenlet.switch() [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] result = function(*args, **kwargs) [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return func(*args, **kwargs) [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise e [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] nwinfo = self.network_api.allocate_for_instance( [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] created_port_ids = self._update_ports_for_instance( [ 632.455189] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] with excutils.save_and_reraise_exception(): [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.force_reraise() [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise self.value [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] updated_port = self._update_port( [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] _ensure_no_port_binding_failure(port) [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise exception.PortBindingFailed(port_id=port['id']) [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 632.455536] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] [ 632.455839] env[62558]: INFO nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Terminating instance [ 632.456417] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.659357] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e253eee-15c3-4a2b-893e-2ee9dac6a7a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.668785] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cb842b-e7e2-4794-85f1-799429fdc314 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.700271] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e0d6ed-8ff0-455e-b092-361d0d11ad69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.708882] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb43b62-e781-4728-8913-26ef413bfef5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.723540] env[62558]: DEBUG nova.compute.provider_tree [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.906808] env[62558]: DEBUG nova.network.neutron [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.024044] env[62558]: DEBUG nova.network.neutron [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.165281] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "2ac801d7-af70-46e0-88b3-02caee13497d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.165281] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.230530] env[62558]: DEBUG nova.scheduler.client.report [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.526594] env[62558]: DEBUG oslo_concurrency.lockutils [req-dabe80be-12a6-46d6-a4fc-19288cd6c951 req-890d2a14-5118-4ff1-87b9-ef3a6423ed78 service nova] Releasing lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.527067] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.527309] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.737440] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.737637] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.740897] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.796s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.742438] env[62558]: INFO nova.compute.claims [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.923758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.924044] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.047943] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.153093] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.254231] env[62558]: DEBUG nova.compute.utils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.255536] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.255836] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.302123] env[62558]: DEBUG nova.policy [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a29309212f34ee88de7e578f7a917a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8f141c2faa24251a9ebe97ede126741', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.435266] env[62558]: DEBUG nova.compute.manager [req-142e5d8b-1f01-404f-8f16-a1198a8d8731 req-469d6df6-24dd-4087-8f01-cb413adfbc21 service nova] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Received event network-vif-deleted-dc2bdb70-ad34-41ce-b709-d7175df751ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.663419] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.663419] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.663419] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.663419] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-259206e4-3acf-453d-a6e0-e77a37f65d57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.673939] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99de9199-1923-4a19-8337-e2d6f90f020a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.699380] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 290e15dc-4ee0-4595-8412-d610895aeb46 could not be found. [ 634.699380] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.699380] env[62558]: INFO nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.699380] env[62558]: DEBUG oslo.service.loopingcall [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.699889] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Successfully created port: c4857d6e-bd1a-42cb-970b-a0c8d920e236 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.701905] env[62558]: DEBUG nova.compute.manager [-] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.702141] env[62558]: DEBUG nova.network.neutron [-] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.719190] env[62558]: DEBUG nova.network.neutron [-] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.759062] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.223101] env[62558]: DEBUG nova.network.neutron [-] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.306367] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9438652-ab38-4998-8336-2c36682a3ed6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.318836] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b434f2-ec9c-4d8e-8d47-daca6ed17df2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.349278] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e83b13-56b0-40ff-bbc4-a396128a65e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.361467] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71ce837-e580-481a-96ef-d3dcaca3d0e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.375810] env[62558]: DEBUG nova.compute.provider_tree [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.727920] env[62558]: INFO nova.compute.manager [-] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Took 1.02 seconds to deallocate network for instance. [ 635.729323] env[62558]: DEBUG nova.compute.claims [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.729323] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.760732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "8acd06fc-b040-468c-980f-6e17b2343c4a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.760732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.768330] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.798081] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.798358] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.798518] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.798695] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.798836] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.798980] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.799202] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.799359] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.799519] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.799674] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.799836] env[62558]: DEBUG nova.virt.hardware [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.800704] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789895cd-17db-488b-9c18-c6894177cb49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.808941] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98ca50d-778b-4a78-b46f-5eeccaeba8ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.879160] env[62558]: DEBUG nova.scheduler.client.report [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.103644] env[62558]: ERROR nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 636.103644] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.103644] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.103644] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.103644] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.103644] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.103644] env[62558]: ERROR nova.compute.manager raise self.value [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.103644] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.103644] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.103644] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.104097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.104097] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.104097] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 636.104097] env[62558]: ERROR nova.compute.manager [ 636.104097] env[62558]: Traceback (most recent call last): [ 636.104097] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.104097] env[62558]: listener.cb(fileno) [ 636.104097] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.104097] env[62558]: result = function(*args, **kwargs) [ 636.104097] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.104097] env[62558]: return func(*args, **kwargs) [ 636.104097] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.104097] env[62558]: raise e [ 636.104097] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.104097] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 636.104097] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.104097] env[62558]: created_port_ids = self._update_ports_for_instance( [ 636.104097] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.104097] env[62558]: with excutils.save_and_reraise_exception(): [ 636.104097] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.104097] env[62558]: self.force_reraise() [ 636.104097] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.104097] env[62558]: raise self.value [ 636.104097] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.104097] env[62558]: updated_port = self._update_port( [ 636.104097] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.104097] env[62558]: _ensure_no_port_binding_failure(port) [ 636.104097] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.104097] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.104845] env[62558]: nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 636.104845] env[62558]: Removing descriptor: 22 [ 636.104845] env[62558]: ERROR nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Traceback (most recent call last): [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] yield resources [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.driver.spawn(context, instance, image_meta, [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.104845] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] vm_ref = self.build_virtual_machine(instance, [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] for vif in network_info: [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self._sync_wrapper(fn, *args, **kwargs) [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.wait() [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self[:] = self._gt.wait() [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self._exit_event.wait() [ 636.105307] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] result = hub.switch() [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self.greenlet.switch() [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] result = function(*args, **kwargs) [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return func(*args, **kwargs) [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise e [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] nwinfo = self.network_api.allocate_for_instance( [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.105685] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] created_port_ids = self._update_ports_for_instance( [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] with excutils.save_and_reraise_exception(): [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.force_reraise() [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise self.value [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] updated_port = self._update_port( [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] _ensure_no_port_binding_failure(port) [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.106086] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise exception.PortBindingFailed(port_id=port['id']) [ 636.106434] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 636.106434] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] [ 636.106434] env[62558]: INFO nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Terminating instance [ 636.107276] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquiring lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.107436] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquired lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.107595] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.383688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.384178] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.387642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.250s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.388851] env[62558]: INFO nova.compute.claims [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.489274] env[62558]: DEBUG nova.compute.manager [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Received event network-changed-c4857d6e-bd1a-42cb-970b-a0c8d920e236 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.489554] env[62558]: DEBUG nova.compute.manager [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Refreshing instance network info cache due to event network-changed-c4857d6e-bd1a-42cb-970b-a0c8d920e236. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 636.489619] env[62558]: DEBUG oslo_concurrency.lockutils [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] Acquiring lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.637658] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.889933] env[62558]: DEBUG nova.compute.utils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.891347] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.891520] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.917644] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.945379] env[62558]: DEBUG nova.policy [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82cfc280dc6a4876822c6c7ce21537f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7594024ac3d24174bf0bf6fcd80e4b65', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.246238] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Successfully created port: f55ea631-b731-494a-9f38-680a3c49831d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.395138] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.422556] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Releasing lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.423016] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 637.423235] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 637.423864] env[62558]: DEBUG oslo_concurrency.lockutils [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] Acquired lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.423864] env[62558]: DEBUG nova.network.neutron [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Refreshing network info cache for port c4857d6e-bd1a-42cb-970b-a0c8d920e236 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.424795] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-719c24b0-6500-4073-85f1-9f7d748d439d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.434940] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139af216-ebbd-44f0-94f6-8e530f706c1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.459317] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5 could not be found. [ 637.459546] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.459722] env[62558]: INFO nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.459957] env[62558]: DEBUG oslo.service.loopingcall [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.462340] env[62558]: DEBUG nova.compute.manager [-] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.462398] env[62558]: DEBUG nova.network.neutron [-] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.484961] env[62558]: DEBUG nova.network.neutron [-] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.892687] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712cd401-6ca3-4a65-a573-45008aecf6d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.900881] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3561702-b81c-4e7b-add9-b2fc8f54b5f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.938582] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565f3c9e-4f17-4386-b4b6-77d1610c002a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.944020] env[62558]: DEBUG nova.compute.manager [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Received event network-changed-f55ea631-b731-494a-9f38-680a3c49831d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.944020] env[62558]: DEBUG nova.compute.manager [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Refreshing instance network info cache due to event network-changed-f55ea631-b731-494a-9f38-680a3c49831d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.944020] env[62558]: DEBUG oslo_concurrency.lockutils [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] Acquiring lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.944020] env[62558]: DEBUG oslo_concurrency.lockutils [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] Acquired lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.944020] env[62558]: DEBUG nova.network.neutron [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Refreshing network info cache for port f55ea631-b731-494a-9f38-680a3c49831d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.952125] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab662cd9-a08f-4eec-952c-9c712c4bb51e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.972966] env[62558]: DEBUG nova.compute.provider_tree [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.975078] env[62558]: DEBUG nova.network.neutron [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.989229] env[62558]: DEBUG nova.network.neutron [-] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.079893] env[62558]: DEBUG nova.network.neutron [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.085381] env[62558]: ERROR nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 638.085381] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.085381] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.085381] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.085381] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.085381] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.085381] env[62558]: ERROR nova.compute.manager raise self.value [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.085381] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.085381] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.085381] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.085795] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.085795] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.085795] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 638.085795] env[62558]: ERROR nova.compute.manager [ 638.085795] env[62558]: Traceback (most recent call last): [ 638.085795] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.085795] env[62558]: listener.cb(fileno) [ 638.085795] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.085795] env[62558]: result = function(*args, **kwargs) [ 638.085795] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.085795] env[62558]: return func(*args, **kwargs) [ 638.085795] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.085795] env[62558]: raise e [ 638.085795] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.085795] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 638.085795] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.085795] env[62558]: created_port_ids = self._update_ports_for_instance( [ 638.085795] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.085795] env[62558]: with excutils.save_and_reraise_exception(): [ 638.085795] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.085795] env[62558]: self.force_reraise() [ 638.085795] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.085795] env[62558]: raise self.value [ 638.085795] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.085795] env[62558]: updated_port = self._update_port( [ 638.085795] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.085795] env[62558]: _ensure_no_port_binding_failure(port) [ 638.085795] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.085795] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.086468] env[62558]: nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 638.086468] env[62558]: Removing descriptor: 15 [ 638.412200] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.436756] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.437014] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.437227] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.437445] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.437593] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.437737] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.437940] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.438108] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.438325] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.438492] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.438662] env[62558]: DEBUG nova.virt.hardware [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.439536] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe06d5a-53f0-4c0f-a76e-5a7752ebb6d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.449456] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3c6582-0563-4815-89a0-9d07974f81fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.464877] env[62558]: ERROR nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Traceback (most recent call last): [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] yield resources [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.driver.spawn(context, instance, image_meta, [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] vm_ref = self.build_virtual_machine(instance, [ 638.464877] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] for vif in network_info: [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return self._sync_wrapper(fn, *args, **kwargs) [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.wait() [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self[:] = self._gt.wait() [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return self._exit_event.wait() [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.465335] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] current.throw(*self._exc) [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] result = function(*args, **kwargs) [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return func(*args, **kwargs) [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise e [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] nwinfo = self.network_api.allocate_for_instance( [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] created_port_ids = self._update_ports_for_instance( [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] with excutils.save_and_reraise_exception(): [ 638.465698] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.force_reraise() [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise self.value [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] updated_port = self._update_port( [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] _ensure_no_port_binding_failure(port) [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise exception.PortBindingFailed(port_id=port['id']) [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 638.466058] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] [ 638.466058] env[62558]: INFO nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Terminating instance [ 638.466934] env[62558]: DEBUG nova.network.neutron [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.468690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquiring lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.479050] env[62558]: DEBUG nova.scheduler.client.report [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.491865] env[62558]: INFO nova.compute.manager [-] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Took 1.03 seconds to deallocate network for instance. [ 638.493945] env[62558]: DEBUG nova.compute.claims [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.494131] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.533869] env[62558]: DEBUG nova.network.neutron [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.582545] env[62558]: DEBUG oslo_concurrency.lockutils [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] Releasing lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.582777] env[62558]: DEBUG nova.compute.manager [req-b4f70fd6-f682-429e-aaf0-c1e91d89778a req-62cc5cb8-c419-41dd-b74e-7bb96a12ba5e service nova] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Received event network-vif-deleted-c4857d6e-bd1a-42cb-970b-a0c8d920e236 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.982765] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.983327] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.985937] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.851s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.986107] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.986260] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 638.986533] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.010s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.990235] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f27e461-9530-4e74-9ae1-cc9897979147 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.998318] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e30288a-c5cf-4199-99f6-2db873675c26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.012477] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd11b6-902b-46e5-b9e4-f424d8fd7ada {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.021693] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29638e8-abb0-491f-9ec5-74ed79c2e0d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.049095] env[62558]: DEBUG oslo_concurrency.lockutils [req-03efd35c-61a9-40a6-9ea5-7bcae785e754 req-d187a3c2-28bc-445c-9bde-e443ac53044e service nova] Releasing lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.049549] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181427MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 639.049655] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.049979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquired lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.050164] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.494226] env[62558]: DEBUG nova.compute.utils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.496596] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.496596] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 639.542399] env[62558]: DEBUG nova.policy [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96ad27112d1b447bbc79a9dbffc0aacb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '253d1165c14f44a09fe13cb085d6c57b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.573491] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.680757] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.792882] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Successfully created port: e792f4d9-3444-4eac-95a7-fd1035bd75f0 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.908551] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92c6865-420b-47f9-8174-19691c1c3af7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.918222] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3972bdd6-be5c-4da6-9c7e-d50b92657f89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.955838] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e24d9a-5f9c-4626-92ff-6f82ba81aa50 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.959548] env[62558]: DEBUG nova.compute.manager [req-2655b106-5ed5-4727-a1a1-aafffbaa5ff5 req-e713e4b2-4c8d-4339-b7ef-3424f70405e9 service nova] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Received event network-vif-deleted-f55ea631-b731-494a-9f38-680a3c49831d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.965533] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b6fa47-3b83-4412-a82d-af67f7e42116 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.979559] env[62558]: DEBUG nova.compute.provider_tree [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.998617] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.183521] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Releasing lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.184057] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.184271] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.184646] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-916d3749-f854-407c-a51e-bad193e03db7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.194203] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8467f7-9ffc-4270-aa5b-a433411bbb4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.215481] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d46bdbc6-e691-44b7-8e9c-3bedb1275e93 could not be found. [ 640.215699] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 640.215878] env[62558]: INFO nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Took 0.03 seconds to destroy the instance on the hypervisor. [ 640.216136] env[62558]: DEBUG oslo.service.loopingcall [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.216373] env[62558]: DEBUG nova.compute.manager [-] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.216480] env[62558]: DEBUG nova.network.neutron [-] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.231230] env[62558]: DEBUG nova.network.neutron [-] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.482605] env[62558]: DEBUG nova.scheduler.client.report [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.641097] env[62558]: ERROR nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 640.641097] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.641097] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.641097] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.641097] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.641097] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.641097] env[62558]: ERROR nova.compute.manager raise self.value [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.641097] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.641097] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.641097] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.641771] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.641771] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.641771] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 640.641771] env[62558]: ERROR nova.compute.manager [ 640.641771] env[62558]: Traceback (most recent call last): [ 640.641771] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.641771] env[62558]: listener.cb(fileno) [ 640.641771] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.641771] env[62558]: result = function(*args, **kwargs) [ 640.641771] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.641771] env[62558]: return func(*args, **kwargs) [ 640.641771] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.641771] env[62558]: raise e [ 640.641771] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.641771] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 640.641771] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.641771] env[62558]: created_port_ids = self._update_ports_for_instance( [ 640.641771] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.641771] env[62558]: with excutils.save_and_reraise_exception(): [ 640.641771] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.641771] env[62558]: self.force_reraise() [ 640.641771] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.641771] env[62558]: raise self.value [ 640.641771] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.641771] env[62558]: updated_port = self._update_port( [ 640.641771] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.641771] env[62558]: _ensure_no_port_binding_failure(port) [ 640.641771] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.641771] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.642583] env[62558]: nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 640.642583] env[62558]: Removing descriptor: 15 [ 640.735020] env[62558]: DEBUG nova.network.neutron [-] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.988659] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.990059] env[62558]: ERROR nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Traceback (most recent call last): [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.driver.spawn(context, instance, image_meta, [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] vm_ref = self.build_virtual_machine(instance, [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.990059] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] for vif in network_info: [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self._sync_wrapper(fn, *args, **kwargs) [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.wait() [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self[:] = self._gt.wait() [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self._exit_event.wait() [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] result = hub.switch() [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.990497] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return self.greenlet.switch() [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] result = function(*args, **kwargs) [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] return func(*args, **kwargs) [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise e [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] nwinfo = self.network_api.allocate_for_instance( [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] created_port_ids = self._update_ports_for_instance( [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] with excutils.save_and_reraise_exception(): [ 640.990912] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] self.force_reraise() [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise self.value [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] updated_port = self._update_port( [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] _ensure_no_port_binding_failure(port) [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] raise exception.PortBindingFailed(port_id=port['id']) [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] nova.exception.PortBindingFailed: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. [ 640.991286] env[62558]: ERROR nova.compute.manager [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] [ 640.991680] env[62558]: DEBUG nova.compute.utils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.991680] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.206s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.992692] env[62558]: INFO nova.compute.claims [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.995304] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Build of instance 2e1410c0-8aff-429c-8bbb-46d9bc01de32 was re-scheduled: Binding failed for port 43fbab1a-a239-4a31-abca-4451749b924d, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.995656] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.995879] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquiring lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.996035] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Acquired lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.996194] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.010420] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.035131] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.035416] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.035575] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.035770] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.035907] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.036063] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.036272] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.036448] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.036635] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.036795] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.036959] env[62558]: DEBUG nova.virt.hardware [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.038088] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7968c464-98d6-43d6-936f-ab31de79e73d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.046660] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af11afb5-dce0-441e-9982-cff5342f2427 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.061367] env[62558]: ERROR nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Traceback (most recent call last): [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] yield resources [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.driver.spawn(context, instance, image_meta, [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] vm_ref = self.build_virtual_machine(instance, [ 641.061367] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] for vif in network_info: [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return self._sync_wrapper(fn, *args, **kwargs) [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.wait() [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self[:] = self._gt.wait() [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return self._exit_event.wait() [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.061881] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] current.throw(*self._exc) [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] result = function(*args, **kwargs) [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return func(*args, **kwargs) [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise e [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] nwinfo = self.network_api.allocate_for_instance( [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] created_port_ids = self._update_ports_for_instance( [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] with excutils.save_and_reraise_exception(): [ 641.062219] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.force_reraise() [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise self.value [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] updated_port = self._update_port( [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] _ensure_no_port_binding_failure(port) [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise exception.PortBindingFailed(port_id=port['id']) [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 641.062585] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] [ 641.062585] env[62558]: INFO nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Terminating instance [ 641.063708] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquiring lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.063863] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquired lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.064034] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.236865] env[62558]: INFO nova.compute.manager [-] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Took 1.02 seconds to deallocate network for instance. [ 641.239473] env[62558]: DEBUG nova.compute.claims [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.239772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.516013] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.568674] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.581631] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.665667] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.973218] env[62558]: DEBUG nova.compute.manager [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Received event network-changed-e792f4d9-3444-4eac-95a7-fd1035bd75f0 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.973434] env[62558]: DEBUG nova.compute.manager [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Refreshing instance network info cache due to event network-changed-e792f4d9-3444-4eac-95a7-fd1035bd75f0. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.973615] env[62558]: DEBUG oslo_concurrency.lockutils [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] Acquiring lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.071425] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Releasing lock "refresh_cache-2e1410c0-8aff-429c-8bbb-46d9bc01de32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.071660] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.071819] env[62558]: DEBUG nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.071983] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.085740] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.168829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Releasing lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.168954] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.169172] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.169636] env[62558]: DEBUG oslo_concurrency.lockutils [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] Acquired lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.170339] env[62558]: DEBUG nova.network.neutron [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Refreshing network info cache for port e792f4d9-3444-4eac-95a7-fd1035bd75f0 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 642.170814] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5b2c97b-6ad5-46db-87f1-6371642eb614 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.188535] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3dc54a-a866-466d-bf61-468593ac86c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.214052] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e7dcc93-2079-4857-8f18-513958ef4e0a could not be found. [ 642.214286] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.214503] env[62558]: INFO nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 642.214697] env[62558]: DEBUG oslo.service.loopingcall [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.217082] env[62558]: DEBUG nova.compute.manager [-] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.217186] env[62558]: DEBUG nova.network.neutron [-] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.236895] env[62558]: DEBUG nova.network.neutron [-] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.531729] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ed6d4f-61eb-47ac-9650-adf88d075476 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.539539] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814ce2e5-98a9-439f-be9a-375067dce3da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.567928] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ba9c15-5bc5-4986-883c-f24fa0d2ab68 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.574658] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f8bacb-b8f9-4be1-b748-3ec02f278ffe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.587181] env[62558]: DEBUG nova.compute.provider_tree [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.588520] env[62558]: DEBUG nova.network.neutron [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.690702] env[62558]: DEBUG nova.network.neutron [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.741169] env[62558]: DEBUG nova.network.neutron [-] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.779102] env[62558]: DEBUG nova.network.neutron [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.090544] env[62558]: DEBUG nova.scheduler.client.report [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.095350] env[62558]: INFO nova.compute.manager [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] [instance: 2e1410c0-8aff-429c-8bbb-46d9bc01de32] Took 1.02 seconds to deallocate network for instance. [ 643.247320] env[62558]: INFO nova.compute.manager [-] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Took 1.03 seconds to deallocate network for instance. [ 643.249665] env[62558]: DEBUG nova.compute.claims [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.249844] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.282053] env[62558]: DEBUG oslo_concurrency.lockutils [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] Releasing lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.282334] env[62558]: DEBUG nova.compute.manager [req-a427d666-4a15-4bfe-bcec-f6a847ca0acb req-72919a2f-cb3d-4c24-97c1-b195b326fd3b service nova] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Received event network-vif-deleted-e792f4d9-3444-4eac-95a7-fd1035bd75f0 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.596214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.598020] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 643.599472] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.783s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.600872] env[62558]: INFO nova.compute.claims [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.109738] env[62558]: DEBUG nova.compute.utils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.114156] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.114228] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 644.131203] env[62558]: INFO nova.scheduler.client.report [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Deleted allocations for instance 2e1410c0-8aff-429c-8bbb-46d9bc01de32 [ 644.172420] env[62558]: DEBUG nova.policy [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d2f0b401df84db29035d06550fb389e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80e71bb1363744a083ca7a5d8bb13cd5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.503755] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Successfully created port: 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.614330] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 644.643424] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aab47662-734d-47f2-a1d9-79700ea793c9 tempest-TenantUsagesTestJSON-1503578095 tempest-TenantUsagesTestJSON-1503578095-project-member] Lock "2e1410c0-8aff-429c-8bbb-46d9bc01de32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.065s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.056096] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcf97c9-a6b2-416b-860a-1f4971fddf4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.064846] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260d3790-8e88-40d6-b1fb-d28a41cc3b27 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.099076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ae8bb8-1ca6-49db-93aa-1c8f7839f37e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.106848] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7454b9-5aca-4f46-af89-94634c522867 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.120370] env[62558]: DEBUG nova.compute.provider_tree [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.149194] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.623936] env[62558]: DEBUG nova.scheduler.client.report [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.629035] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 645.638383] env[62558]: DEBUG nova.compute.manager [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Received event network-changed-28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.638383] env[62558]: DEBUG nova.compute.manager [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Refreshing instance network info cache due to event network-changed-28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 645.638383] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] Acquiring lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.638383] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] Acquired lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.638383] env[62558]: DEBUG nova.network.neutron [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Refreshing network info cache for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 645.670146] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.670146] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.670146] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.670302] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.670302] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.670302] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.670302] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.670302] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.670469] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.670787] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.671146] env[62558]: DEBUG nova.virt.hardware [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.675457] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde0d0cd-bac6-472e-b2ed-74b7c42bb6c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.680694] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.687758] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d7d5fa-a06e-404e-b1d0-30a0a4a6f826 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.991105] env[62558]: ERROR nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 645.991105] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.991105] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.991105] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.991105] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.991105] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.991105] env[62558]: ERROR nova.compute.manager raise self.value [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.991105] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.991105] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.991105] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.991498] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.991498] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.991498] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 645.991498] env[62558]: ERROR nova.compute.manager [ 645.991498] env[62558]: Traceback (most recent call last): [ 645.991498] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.991498] env[62558]: listener.cb(fileno) [ 645.991498] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.991498] env[62558]: result = function(*args, **kwargs) [ 645.991498] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.991498] env[62558]: return func(*args, **kwargs) [ 645.991498] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.991498] env[62558]: raise e [ 645.991498] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.991498] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 645.991498] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.991498] env[62558]: created_port_ids = self._update_ports_for_instance( [ 645.991498] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.991498] env[62558]: with excutils.save_and_reraise_exception(): [ 645.991498] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.991498] env[62558]: self.force_reraise() [ 645.991498] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.991498] env[62558]: raise self.value [ 645.991498] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.991498] env[62558]: updated_port = self._update_port( [ 645.991498] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.991498] env[62558]: _ensure_no_port_binding_failure(port) [ 645.991498] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.991498] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.992165] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 645.992165] env[62558]: Removing descriptor: 22 [ 645.992165] env[62558]: ERROR nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Traceback (most recent call last): [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] yield resources [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.driver.spawn(context, instance, image_meta, [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.992165] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] vm_ref = self.build_virtual_machine(instance, [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] for vif in network_info: [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self._sync_wrapper(fn, *args, **kwargs) [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.wait() [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self[:] = self._gt.wait() [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self._exit_event.wait() [ 645.992456] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] result = hub.switch() [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self.greenlet.switch() [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] result = function(*args, **kwargs) [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return func(*args, **kwargs) [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise e [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] nwinfo = self.network_api.allocate_for_instance( [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.992805] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] created_port_ids = self._update_ports_for_instance( [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] with excutils.save_and_reraise_exception(): [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.force_reraise() [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise self.value [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] updated_port = self._update_port( [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] _ensure_no_port_binding_failure(port) [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.993402] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise exception.PortBindingFailed(port_id=port['id']) [ 645.993773] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 645.993773] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] [ 645.993773] env[62558]: INFO nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Terminating instance [ 645.994396] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquiring lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.132888] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.133448] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.136377] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.265s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.137859] env[62558]: INFO nova.compute.claims [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.155694] env[62558]: DEBUG nova.network.neutron [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.281381] env[62558]: DEBUG nova.network.neutron [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.730029] env[62558]: DEBUG nova.compute.utils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.730029] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 646.730029] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 646.730029] env[62558]: DEBUG nova.policy [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da1f27f9a6ac44a0a2dc286b7f0888a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fabbd2127c2445aa96cf0d513541d32', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.784262] env[62558]: DEBUG oslo_concurrency.lockutils [req-5c2c617e-5033-445f-b90b-88f50cbf6701 req-83245959-112f-4ac6-9081-ec5bd1622aa1 service nova] Releasing lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.784262] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquired lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.784561] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.027210] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Successfully created port: dd165904-c6de-4b98-8adf-d2d1308a20aa {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.153083] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.305873] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.396090] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.607076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37af88aa-1962-4bd5-89aa-0fb2d67e3ea8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.616104] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d502036-9cbc-4632-8d87-249321fe783a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.653731] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5536dedd-9751-4883-a26f-605b433ddf89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.661432] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a669d72-f625-4af1-bde3-1bbb4f86f17f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.680613] env[62558]: DEBUG nova.compute.provider_tree [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.840170] env[62558]: DEBUG nova.compute.manager [req-c85dc94a-fe07-479e-bc90-dc7439b26fe7 req-2f6e23fb-7aeb-4ba2-88fb-2c24f08eb6d3 service nova] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Received event network-vif-deleted-28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.900767] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Releasing lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.900767] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 647.900767] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.901017] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c08a43a8-3f61-46aa-a847-69e305d8d433 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.913401] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35431e17-d873-43ee-942f-2628a7124a53 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.937020] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9274ce52-49af-48df-bcd1-269b8241d93f could not be found. [ 647.937267] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.937464] env[62558]: INFO nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 647.937863] env[62558]: DEBUG oslo.service.loopingcall [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 647.938057] env[62558]: DEBUG nova.compute.manager [-] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.938151] env[62558]: DEBUG nova.network.neutron [-] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.954144] env[62558]: DEBUG nova.network.neutron [-] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.172904] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.183686] env[62558]: DEBUG nova.scheduler.client.report [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.203121] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.203431] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.203638] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.203853] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.204046] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.204230] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.204486] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.204725] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.204960] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.205174] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.205545] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.206929] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f10f2a7-2b0f-4a3f-ac7b-9482a171dd2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.220109] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652428d3-25da-4ac8-9f4c-77c06ec801ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.456904] env[62558]: DEBUG nova.network.neutron [-] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.486789] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 648.486789] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.486789] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.486789] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.486789] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.486789] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.486789] env[62558]: ERROR nova.compute.manager raise self.value [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.486789] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.486789] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.486789] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.487209] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.487209] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.487209] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 648.487209] env[62558]: ERROR nova.compute.manager [ 648.487209] env[62558]: Traceback (most recent call last): [ 648.487209] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.487209] env[62558]: listener.cb(fileno) [ 648.487209] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.487209] env[62558]: result = function(*args, **kwargs) [ 648.487209] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.487209] env[62558]: return func(*args, **kwargs) [ 648.487209] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.487209] env[62558]: raise e [ 648.487209] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.487209] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 648.487209] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.487209] env[62558]: created_port_ids = self._update_ports_for_instance( [ 648.487209] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.487209] env[62558]: with excutils.save_and_reraise_exception(): [ 648.487209] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.487209] env[62558]: self.force_reraise() [ 648.487209] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.487209] env[62558]: raise self.value [ 648.487209] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.487209] env[62558]: updated_port = self._update_port( [ 648.487209] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.487209] env[62558]: _ensure_no_port_binding_failure(port) [ 648.487209] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.487209] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.488078] env[62558]: nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 648.488078] env[62558]: Removing descriptor: 22 [ 648.488078] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Traceback (most recent call last): [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] yield resources [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.driver.spawn(context, instance, image_meta, [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.488078] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] vm_ref = self.build_virtual_machine(instance, [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] for vif in network_info: [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self._sync_wrapper(fn, *args, **kwargs) [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.wait() [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self[:] = self._gt.wait() [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self._exit_event.wait() [ 648.488444] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] result = hub.switch() [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self.greenlet.switch() [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] result = function(*args, **kwargs) [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return func(*args, **kwargs) [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise e [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] nwinfo = self.network_api.allocate_for_instance( [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.488764] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] created_port_ids = self._update_ports_for_instance( [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] with excutils.save_and_reraise_exception(): [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.force_reraise() [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise self.value [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] updated_port = self._update_port( [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] _ensure_no_port_binding_failure(port) [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.489491] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise exception.PortBindingFailed(port_id=port['id']) [ 648.489829] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 648.489829] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] [ 648.489829] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Terminating instance [ 648.490617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.490873] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.491097] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.691355] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.692041] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 648.694848] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.493s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.961889] env[62558]: INFO nova.compute.manager [-] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Took 1.02 seconds to deallocate network for instance. [ 648.968304] env[62558]: DEBUG nova.compute.claims [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 648.968304] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.008867] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.089603] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.203346] env[62558]: DEBUG nova.compute.utils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.204810] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 649.204978] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 649.246925] env[62558]: DEBUG nova.policy [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da1f27f9a6ac44a0a2dc286b7f0888a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fabbd2127c2445aa96cf0d513541d32', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 649.594073] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.594073] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 649.594251] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.595762] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca56ffe-a7f6-45ea-a995-e1a91730a0d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.608439] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f463410-ef8e-42c3-a616-89a09679980b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.635311] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 88cee397-bcbd-428c-b203-e5d9b6acbbb2 could not be found. [ 649.635581] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 649.635856] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 649.636628] env[62558]: DEBUG oslo.service.loopingcall [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.636628] env[62558]: DEBUG nova.compute.manager [-] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.636628] env[62558]: DEBUG nova.network.neutron [-] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.660141] env[62558]: DEBUG nova.network.neutron [-] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.662714] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Successfully created port: 23f5e7bb-7d33-4809-b60a-9dc4e19a6549 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.692192] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ca99c4-de4d-4f67-82b8-e281e8daeae6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.699982] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bb47f4-23f0-4726-a74b-9df03746820e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.732765] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.736957] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0e5df6-330b-495f-9f1e-651151503992 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.744529] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d457bf2c-aed8-415e-b771-75ca8b6bd31b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.758035] env[62558]: DEBUG nova.compute.provider_tree [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.869372] env[62558]: DEBUG nova.compute.manager [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Received event network-changed-dd165904-c6de-4b98-8adf-d2d1308a20aa {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.869573] env[62558]: DEBUG nova.compute.manager [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Refreshing instance network info cache due to event network-changed-dd165904-c6de-4b98-8adf-d2d1308a20aa. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 649.870097] env[62558]: DEBUG oslo_concurrency.lockutils [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] Acquiring lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.870260] env[62558]: DEBUG oslo_concurrency.lockutils [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] Acquired lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.870428] env[62558]: DEBUG nova.network.neutron [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Refreshing network info cache for port dd165904-c6de-4b98-8adf-d2d1308a20aa {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.164029] env[62558]: DEBUG nova.network.neutron [-] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.263122] env[62558]: DEBUG nova.scheduler.client.report [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.393308] env[62558]: DEBUG nova.network.neutron [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.478614] env[62558]: DEBUG nova.network.neutron [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.667599] env[62558]: INFO nova.compute.manager [-] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Took 1.03 seconds to deallocate network for instance. [ 650.671097] env[62558]: DEBUG nova.compute.claims [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.671555] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.743589] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.771785] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.077s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.772444] env[62558]: ERROR nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] Traceback (most recent call last): [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.driver.spawn(context, instance, image_meta, [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] vm_ref = self.build_virtual_machine(instance, [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.772444] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] for vif in network_info: [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self._sync_wrapper(fn, *args, **kwargs) [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.wait() [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self[:] = self._gt.wait() [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self._exit_event.wait() [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] result = hub.switch() [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.772775] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return self.greenlet.switch() [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] result = function(*args, **kwargs) [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] return func(*args, **kwargs) [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise e [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] nwinfo = self.network_api.allocate_for_instance( [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] created_port_ids = self._update_ports_for_instance( [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] with excutils.save_and_reraise_exception(): [ 650.773156] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] self.force_reraise() [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise self.value [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] updated_port = self._update_port( [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] _ensure_no_port_binding_failure(port) [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] raise exception.PortBindingFailed(port_id=port['id']) [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] nova.exception.PortBindingFailed: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. [ 650.773634] env[62558]: ERROR nova.compute.manager [instance: 95348a45-f517-4e21-8969-57c978985984] [ 650.773949] env[62558]: DEBUG nova.compute.utils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.774268] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.385s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.775777] env[62558]: INFO nova.compute.claims [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.778500] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Build of instance 95348a45-f517-4e21-8969-57c978985984 was re-scheduled: Binding failed for port 15f5b8c8-8fc2-4d2c-984e-9302eb332106, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.778938] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.779188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquiring lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.779334] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Acquired lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.779490] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.783184] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.783184] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.783184] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.783432] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.783583] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.783724] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.783919] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.784087] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.784247] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.784404] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.784567] env[62558]: DEBUG nova.virt.hardware [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.785544] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ee40c5-350d-44ae-be4d-aade66c19a15 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.798331] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db73bd5e-7c01-4b25-984d-b0b9e01f024a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.815165] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.909064] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.984255] env[62558]: DEBUG oslo_concurrency.lockutils [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] Releasing lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.984676] env[62558]: DEBUG nova.compute.manager [req-60a57913-35f2-461f-b69c-ce5d0fcdfdd5 req-ba583237-1275-4f67-8eb2-d330892e9125 service nova] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Received event network-vif-deleted-dd165904-c6de-4b98-8adf-d2d1308a20aa {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.245049] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.245234] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.256378] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 651.256378] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.256378] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.256378] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.256378] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.256378] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.256378] env[62558]: ERROR nova.compute.manager raise self.value [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.256378] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.256378] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.256378] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.256942] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.256942] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.256942] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 651.256942] env[62558]: ERROR nova.compute.manager [ 651.256942] env[62558]: Traceback (most recent call last): [ 651.256942] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.256942] env[62558]: listener.cb(fileno) [ 651.256942] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.256942] env[62558]: result = function(*args, **kwargs) [ 651.256942] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.256942] env[62558]: return func(*args, **kwargs) [ 651.256942] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.256942] env[62558]: raise e [ 651.256942] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.256942] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 651.256942] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.256942] env[62558]: created_port_ids = self._update_ports_for_instance( [ 651.256942] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.256942] env[62558]: with excutils.save_and_reraise_exception(): [ 651.256942] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.256942] env[62558]: self.force_reraise() [ 651.256942] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.256942] env[62558]: raise self.value [ 651.256942] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.256942] env[62558]: updated_port = self._update_port( [ 651.256942] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.256942] env[62558]: _ensure_no_port_binding_failure(port) [ 651.256942] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.256942] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.257926] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 651.257926] env[62558]: Removing descriptor: 22 [ 651.257926] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] Traceback (most recent call last): [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] yield resources [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.driver.spawn(context, instance, image_meta, [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.257926] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] vm_ref = self.build_virtual_machine(instance, [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] for vif in network_info: [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self._sync_wrapper(fn, *args, **kwargs) [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.wait() [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self[:] = self._gt.wait() [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self._exit_event.wait() [ 651.258371] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] result = hub.switch() [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self.greenlet.switch() [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] result = function(*args, **kwargs) [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return func(*args, **kwargs) [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise e [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] nwinfo = self.network_api.allocate_for_instance( [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.258830] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] created_port_ids = self._update_ports_for_instance( [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] with excutils.save_and_reraise_exception(): [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.force_reraise() [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise self.value [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] updated_port = self._update_port( [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] _ensure_no_port_binding_failure(port) [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.259415] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise exception.PortBindingFailed(port_id=port['id']) [ 651.259880] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 651.259880] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] [ 651.259880] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Terminating instance [ 651.260997] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.261238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.261556] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.412401] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Releasing lock "refresh_cache-95348a45-f517-4e21-8969-57c978985984" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.412401] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.412401] env[62558]: DEBUG nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.412401] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.437365] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.787432] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.901093] env[62558]: DEBUG nova.compute.manager [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Received event network-changed-23f5e7bb-7d33-4809-b60a-9dc4e19a6549 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.901306] env[62558]: DEBUG nova.compute.manager [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Refreshing instance network info cache due to event network-changed-23f5e7bb-7d33-4809-b60a-9dc4e19a6549. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.901529] env[62558]: DEBUG oslo_concurrency.lockutils [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] Acquiring lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.913705] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.940435] env[62558]: DEBUG nova.network.neutron [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.253213] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e597098-96f1-4f59-adae-d6b3267ec5ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.261275] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889a58d5-23b5-4eaf-917e-9c12e4e78d9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.297169] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bd6824-376c-44f1-964d-ffe46ebc479e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.306027] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7173aa3-66bb-4002-9f75-feba1988a02a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.321425] env[62558]: DEBUG nova.compute.provider_tree [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.417189] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.417727] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.417923] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.418286] env[62558]: DEBUG oslo_concurrency.lockutils [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] Acquired lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.418518] env[62558]: DEBUG nova.network.neutron [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Refreshing network info cache for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.420023] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d993f3c4-6cde-48d6-a8e1-d57892fde5a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.429570] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa22ef7-ba53-40d9-a939-9ca3d42edbbd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.443708] env[62558]: INFO nova.compute.manager [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] [instance: 95348a45-f517-4e21-8969-57c978985984] Took 1.03 seconds to deallocate network for instance. [ 652.451923] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef19262f-77be-470a-b510-29b8016da5a5 could not be found. [ 652.452161] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.452318] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 652.452561] env[62558]: DEBUG oslo.service.loopingcall [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.453034] env[62558]: DEBUG nova.compute.manager [-] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.453137] env[62558]: DEBUG nova.network.neutron [-] [instance: ef19262f-77be-470a-b510-29b8016da5a5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.480010] env[62558]: DEBUG nova.network.neutron [-] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.824595] env[62558]: DEBUG nova.scheduler.client.report [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.940570] env[62558]: DEBUG nova.network.neutron [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.981567] env[62558]: DEBUG nova.network.neutron [-] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.020265] env[62558]: DEBUG nova.network.neutron [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.329491] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.330066] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.332708] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.604s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.475314] env[62558]: INFO nova.scheduler.client.report [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Deleted allocations for instance 95348a45-f517-4e21-8969-57c978985984 [ 653.486370] env[62558]: INFO nova.compute.manager [-] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Took 1.03 seconds to deallocate network for instance. [ 653.488248] env[62558]: DEBUG nova.compute.claims [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.488415] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.522700] env[62558]: DEBUG oslo_concurrency.lockutils [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] Releasing lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.523555] env[62558]: DEBUG nova.compute.manager [req-2f893bd2-48a1-4621-970b-67532301f2ed req-72b72e13-ebcf-4e77-b768-2247aadba794 service nova] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Received event network-vif-deleted-23f5e7bb-7d33-4809-b60a-9dc4e19a6549 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.837259] env[62558]: DEBUG nova.compute.utils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.838787] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.838903] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.915528] env[62558]: DEBUG nova.policy [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e3869eb3be44f2390047e18ee9cca06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5af210e786cc43018ea55783e5961cfa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.984399] env[62558]: DEBUG oslo_concurrency.lockutils [None req-77e3e473-2f9d-423a-bdf9-a058f2cef9ba tempest-ImagesNegativeTestJSON-611852905 tempest-ImagesNegativeTestJSON-611852905-project-member] Lock "95348a45-f517-4e21-8969-57c978985984" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.626s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.282784] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef6ed6e-2d6a-4f33-9973-69583aeb0fd1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.291907] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf21e83-8bc6-40ab-a3a4-218eb706ee9b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.323089] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfec6b7-bdfa-442d-ae79-b29cde91577d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.330887] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d96bde9-c14e-4ae0-a4e8-fe61c2743a4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.337142] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Successfully created port: b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.346978] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.349903] env[62558]: DEBUG nova.compute.provider_tree [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.493864] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.861758] env[62558]: DEBUG nova.scheduler.client.report [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.016587] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.372023] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.374281] env[62558]: DEBUG nova.compute.manager [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Received event network-changed-b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.374454] env[62558]: DEBUG nova.compute.manager [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Refreshing instance network info cache due to event network-changed-b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 655.374844] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] Acquiring lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.374844] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] Acquired lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.376334] env[62558]: DEBUG nova.network.neutron [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Refreshing network info cache for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.376334] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.377019] env[62558]: ERROR nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Traceback (most recent call last): [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.driver.spawn(context, instance, image_meta, [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] vm_ref = self.build_virtual_machine(instance, [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.377019] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] for vif in network_info: [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self._sync_wrapper(fn, *args, **kwargs) [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.wait() [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self[:] = self._gt.wait() [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self._exit_event.wait() [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] result = hub.switch() [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.379145] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return self.greenlet.switch() [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] result = function(*args, **kwargs) [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] return func(*args, **kwargs) [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise e [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] nwinfo = self.network_api.allocate_for_instance( [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] created_port_ids = self._update_ports_for_instance( [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] with excutils.save_and_reraise_exception(): [ 655.381158] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] self.force_reraise() [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise self.value [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] updated_port = self._update_port( [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] _ensure_no_port_binding_failure(port) [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] raise exception.PortBindingFailed(port_id=port['id']) [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] nova.exception.PortBindingFailed: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. [ 655.383574] env[62558]: ERROR nova.compute.manager [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] [ 655.383843] env[62558]: DEBUG nova.compute.utils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.383843] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.885s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.383843] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Build of instance 290e15dc-4ee0-4595-8412-d610895aeb46 was re-scheduled: Binding failed for port dc2bdb70-ad34-41ce-b709-d7175df751ec, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.383843] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.389731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.389731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.389731] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.419057] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.419220] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.419377] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.419552] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.419698] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.419838] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.422709] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.422848] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.423304] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.423304] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.425110] env[62558]: DEBUG nova.virt.hardware [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.425110] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7cb89a-bb1e-486e-a90a-7df79023f8cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.433359] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4756c6d-b63c-403d-8305-0eb6e81043c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.573729] env[62558]: ERROR nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 655.573729] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.573729] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.573729] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.573729] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.573729] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.573729] env[62558]: ERROR nova.compute.manager raise self.value [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.573729] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.573729] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.573729] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.574609] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.574609] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.574609] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 655.574609] env[62558]: ERROR nova.compute.manager [ 655.574609] env[62558]: Traceback (most recent call last): [ 655.574609] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.574609] env[62558]: listener.cb(fileno) [ 655.574609] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.574609] env[62558]: result = function(*args, **kwargs) [ 655.574609] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.574609] env[62558]: return func(*args, **kwargs) [ 655.574609] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.574609] env[62558]: raise e [ 655.574609] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.574609] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 655.574609] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.574609] env[62558]: created_port_ids = self._update_ports_for_instance( [ 655.574609] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.574609] env[62558]: with excutils.save_and_reraise_exception(): [ 655.574609] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.574609] env[62558]: self.force_reraise() [ 655.574609] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.574609] env[62558]: raise self.value [ 655.574609] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.574609] env[62558]: updated_port = self._update_port( [ 655.574609] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.574609] env[62558]: _ensure_no_port_binding_failure(port) [ 655.574609] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.574609] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.575458] env[62558]: nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 655.575458] env[62558]: Removing descriptor: 22 [ 655.575458] env[62558]: ERROR nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Traceback (most recent call last): [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] yield resources [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.driver.spawn(context, instance, image_meta, [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.575458] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] vm_ref = self.build_virtual_machine(instance, [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] for vif in network_info: [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self._sync_wrapper(fn, *args, **kwargs) [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.wait() [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self[:] = self._gt.wait() [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self._exit_event.wait() [ 655.575881] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] result = hub.switch() [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self.greenlet.switch() [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] result = function(*args, **kwargs) [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return func(*args, **kwargs) [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise e [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] nwinfo = self.network_api.allocate_for_instance( [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.576283] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] created_port_ids = self._update_ports_for_instance( [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] with excutils.save_and_reraise_exception(): [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.force_reraise() [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise self.value [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] updated_port = self._update_port( [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] _ensure_no_port_binding_failure(port) [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.576607] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise exception.PortBindingFailed(port_id=port['id']) [ 655.576998] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 655.576998] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] [ 655.576998] env[62558]: INFO nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Terminating instance [ 655.580217] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquiring lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.896952] env[62558]: DEBUG nova.network.neutron [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.908442] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.989991] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.000496] env[62558]: DEBUG nova.network.neutron [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.281445] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3c27d5-c579-41f5-9696-18409709ac29 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.290182] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caae042b-5782-492a-97ed-a5cf45ff2961 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.320586] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4676c3-65db-4fe9-b455-153229ce3f30 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.327795] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1359e438-bc72-4487-8397-0a56495853bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.341098] env[62558]: DEBUG nova.compute.provider_tree [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.495324] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-290e15dc-4ee0-4595-8412-d610895aeb46" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.495575] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.495892] env[62558]: DEBUG nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.495965] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.502262] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d71ae5b-39c1-4236-8d1b-aa4cf0e4dd94 req-4aeea447-7a4f-439e-b7d9-3ca08f7f5a2f service nova] Releasing lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.502635] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquired lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.502788] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.511143] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.846123] env[62558]: DEBUG nova.scheduler.client.report [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.013864] env[62558]: DEBUG nova.network.neutron [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.027212] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.136978] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.351487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.352176] env[62558]: ERROR nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Traceback (most recent call last): [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.driver.spawn(context, instance, image_meta, [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] vm_ref = self.build_virtual_machine(instance, [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.352176] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] for vif in network_info: [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self._sync_wrapper(fn, *args, **kwargs) [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.wait() [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self[:] = self._gt.wait() [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self._exit_event.wait() [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] result = hub.switch() [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.352478] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return self.greenlet.switch() [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] result = function(*args, **kwargs) [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] return func(*args, **kwargs) [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise e [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] nwinfo = self.network_api.allocate_for_instance( [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] created_port_ids = self._update_ports_for_instance( [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] with excutils.save_and_reraise_exception(): [ 657.352791] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] self.force_reraise() [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise self.value [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] updated_port = self._update_port( [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] _ensure_no_port_binding_failure(port) [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] raise exception.PortBindingFailed(port_id=port['id']) [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] nova.exception.PortBindingFailed: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. [ 657.353107] env[62558]: ERROR nova.compute.manager [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] [ 657.353411] env[62558]: DEBUG nova.compute.utils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.356351] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.307s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.362018] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Build of instance 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5 was re-scheduled: Binding failed for port c4857d6e-bd1a-42cb-970b-a0c8d920e236, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.362018] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.362018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquiring lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.362018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Acquired lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.362240] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.436967] env[62558]: DEBUG nova.compute.manager [req-5934096c-812b-42fd-882f-cce69e51b36b req-e1b844b4-2231-47af-8100-2b6cdb15d342 service nova] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Received event network-vif-deleted-b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.518698] env[62558]: INFO nova.compute.manager [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 290e15dc-4ee0-4595-8412-d610895aeb46] Took 1.02 seconds to deallocate network for instance. [ 657.640110] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Releasing lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.640110] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 657.640246] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.641918] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-228d66e9-c784-412e-b066-af71770d5247 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.653443] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b21f258-0f21-4fbc-ac48-5f61a0dbc259 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.676632] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2236b710-2141-49a2-b3f5-9ed71090e1db could not be found. [ 657.676938] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.677180] env[62558]: INFO nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Took 0.04 seconds to destroy the instance on the hypervisor. [ 657.677468] env[62558]: DEBUG oslo.service.loopingcall [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.677762] env[62558]: DEBUG nova.compute.manager [-] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.677928] env[62558]: DEBUG nova.network.neutron [-] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.701532] env[62558]: DEBUG nova.network.neutron [-] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.889807] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.999748] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.203237] env[62558]: DEBUG nova.network.neutron [-] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.503121] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Releasing lock "refresh_cache-49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.503415] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.503492] env[62558]: DEBUG nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.504648] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.524387] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.550202] env[62558]: INFO nova.scheduler.client.report [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocations for instance 290e15dc-4ee0-4595-8412-d610895aeb46 [ 658.705541] env[62558]: INFO nova.compute.manager [-] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Took 1.03 seconds to deallocate network for instance. [ 658.707969] env[62558]: DEBUG nova.compute.claims [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.708165] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.913423] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 290e15dc-4ee0-4595-8412-d610895aeb46 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.027029] env[62558]: DEBUG nova.network.neutron [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.066899] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a0f4acc1-ff5f-4ec8-9c60-79e263b93f0d tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "290e15dc-4ee0-4595-8412-d610895aeb46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.081s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.107581] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.107850] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.416620] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.416791] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d46bdbc6-e691-44b7-8e9c-3bedb1275e93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.416983] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 8e7dcc93-2079-4857-8f18-513958ef4e0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.417229] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9274ce52-49af-48df-bcd1-269b8241d93f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.417382] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 88cee397-bcbd-428c-b203-e5d9b6acbbb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.417505] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ef19262f-77be-470a-b510-29b8016da5a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.417661] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2236b710-2141-49a2-b3f5-9ed71090e1db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 659.530056] env[62558]: INFO nova.compute.manager [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] [instance: 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5] Took 1.03 seconds to deallocate network for instance. [ 659.571267] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.920385] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance dca0a819-d11d-437e-9a4c-da4279a9d553 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.095402] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.141537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "6d89107f-0727-4b8c-863d-d1e635000bff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.141770] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.423347] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 65e8971a-a149-4a11-92ad-7c5c8874530c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.561058] env[62558]: INFO nova.scheduler.client.report [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Deleted allocations for instance 49ba74b6-fbbe-442a-b3a1-00bdd4e046e5 [ 660.927041] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance b482dcca-ae32-425d-a097-f9af69785835 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.070494] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f46fc6d6-7ed4-4c2a-9ad7-f155d19ee34a tempest-ServerPasswordTestJSON-784271367 tempest-ServerPasswordTestJSON-784271367-project-member] Lock "49ba74b6-fbbe-442a-b3a1-00bdd4e046e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.888s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.430116] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f8db751a-c495-467c-ae32-addb4387346d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.573258] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.932706] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9b49513e-4c01-4664-be2c-6058a357b0db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.099872] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.436946] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.940564] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2888fde6-8a1b-41ce-a482-cb99acd74a6e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.443817] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ae701c0-3016-4635-8fe1-1e1c314fa748 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.946834] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance cecebc6a-25cc-4376-8467-2aa0f909c9bd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.450254] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a092fb50-c750-4b91-a94e-e11e9eb7faf5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.953708] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 151768fd-c2c8-491f-93b5-bbc53bd69e2c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.456938] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f3128303-9b2c-4f14-9ea0-093f55ecb063 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.960883] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 3e035ee1-a63a-46a6-84ca-4980e1185e03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.464081] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d526aa96-4007-45f8-8eee-9030bdb9f402 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.966824] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d92dca09-0a42-4103-ac92-25b6b0fa3c32 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.470858] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.976510] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d65fbfe3-b1e9-48d9-88ba-209fdc264283 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.479961] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.983363] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6e01c855-ef25-45c9-a73e-5656e8e05cd4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.487494] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance b0038711-5d12-4909-b331-72acb5ed0d24 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.991036] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ac801d7-af70-46e0-88b3-02caee13497d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.493700] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.997097] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 8acd06fc-b040-468c-980f-6e17b2343c4a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.500014] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 3c8c2d25-981e-49de-b5c0-e0d12dd91378 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.500396] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 671.500541] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 671.864699] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11aa78c2-9697-4509-b33b-195ce8cda9c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.872644] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93774bba-3a48-48fd-92c4-6ead124d43a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.901913] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d72acbe-7fa4-4beb-870a-4e6b66941921 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.909172] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a179cbb2-6488-4bde-967d-c65dce5d47b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.922780] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.427013] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.932954] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 672.933315] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.577s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.933772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.694s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.765677] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ffa405-5f71-4302-bf46-f756818013e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.773349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b0d379-4361-4b23-9489-877f61afdf0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.801902] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351fa93c-0455-4798-af83-681bb872e723 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.809020] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cd52cf-5732-4e7d-907c-0162e7bbb81c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.821795] env[62558]: DEBUG nova.compute.provider_tree [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.324785] env[62558]: DEBUG nova.scheduler.client.report [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.829809] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.830519] env[62558]: ERROR nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Traceback (most recent call last): [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.driver.spawn(context, instance, image_meta, [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] vm_ref = self.build_virtual_machine(instance, [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.830519] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] for vif in network_info: [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return self._sync_wrapper(fn, *args, **kwargs) [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.wait() [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self[:] = self._gt.wait() [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return self._exit_event.wait() [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] current.throw(*self._exc) [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.830833] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] result = function(*args, **kwargs) [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] return func(*args, **kwargs) [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise e [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] nwinfo = self.network_api.allocate_for_instance( [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] created_port_ids = self._update_ports_for_instance( [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] with excutils.save_and_reraise_exception(): [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] self.force_reraise() [ 674.831160] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise self.value [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] updated_port = self._update_port( [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] _ensure_no_port_binding_failure(port) [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] raise exception.PortBindingFailed(port_id=port['id']) [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] nova.exception.PortBindingFailed: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. [ 674.831475] env[62558]: ERROR nova.compute.manager [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] [ 674.831475] env[62558]: DEBUG nova.compute.utils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.832622] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.583s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.835565] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Build of instance d46bdbc6-e691-44b7-8e9c-3bedb1275e93 was re-scheduled: Binding failed for port f55ea631-b731-494a-9f38-680a3c49831d, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.835982] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.836229] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquiring lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.836371] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Acquired lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.836529] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.356528] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.478635] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.687138] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0269eed-44c4-4c2f-bb0b-5fe3753f7f12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.694906] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c1cc10-f43a-4e1c-9987-8b4e92c71af7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.723674] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20aec6cc-717d-498f-86c0-978129e6c395 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.730545] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969b6849-4b24-461e-9901-c306dcd6c5a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.742993] env[62558]: DEBUG nova.compute.provider_tree [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.982172] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Releasing lock "refresh_cache-d46bdbc6-e691-44b7-8e9c-3bedb1275e93" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.982457] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.982606] env[62558]: DEBUG nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.982792] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.997301] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.245497] env[62558]: DEBUG nova.scheduler.client.report [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.499796] env[62558]: DEBUG nova.network.neutron [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.750380] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.751052] env[62558]: ERROR nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Traceback (most recent call last): [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.driver.spawn(context, instance, image_meta, [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] vm_ref = self.build_virtual_machine(instance, [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.751052] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] for vif in network_info: [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return self._sync_wrapper(fn, *args, **kwargs) [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.wait() [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self[:] = self._gt.wait() [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return self._exit_event.wait() [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] current.throw(*self._exc) [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.751370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] result = function(*args, **kwargs) [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] return func(*args, **kwargs) [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise e [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] nwinfo = self.network_api.allocate_for_instance( [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] created_port_ids = self._update_ports_for_instance( [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] with excutils.save_and_reraise_exception(): [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] self.force_reraise() [ 676.751689] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise self.value [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] updated_port = self._update_port( [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] _ensure_no_port_binding_failure(port) [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] raise exception.PortBindingFailed(port_id=port['id']) [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] nova.exception.PortBindingFailed: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. [ 676.752370] env[62558]: ERROR nova.compute.manager [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] [ 676.752370] env[62558]: DEBUG nova.compute.utils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.752998] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.072s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.754820] env[62558]: INFO nova.compute.claims [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.757554] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Build of instance 8e7dcc93-2079-4857-8f18-513958ef4e0a was re-scheduled: Binding failed for port e792f4d9-3444-4eac-95a7-fd1035bd75f0, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.758073] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.758372] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquiring lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.758527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Acquired lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.758693] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.003119] env[62558]: INFO nova.compute.manager [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] [instance: d46bdbc6-e691-44b7-8e9c-3bedb1275e93] Took 1.02 seconds to deallocate network for instance. [ 677.284858] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.360897] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.863577] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Releasing lock "refresh_cache-8e7dcc93-2079-4857-8f18-513958ef4e0a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.863810] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.863981] env[62558]: DEBUG nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.864163] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.878107] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.030566] env[62558]: INFO nova.scheduler.client.report [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Deleted allocations for instance d46bdbc6-e691-44b7-8e9c-3bedb1275e93 [ 678.111462] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1519451c-9a7d-456f-b9b3-10d67b165bf6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.119284] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf428e-6c4d-4749-b5d9-995d7c4db20e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.149279] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55878636-1fee-4867-9cd6-e9a25a8da066 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.156440] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc439370-fc1d-4e2e-ae9f-1c8130417876 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.169276] env[62558]: DEBUG nova.compute.provider_tree [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.380481] env[62558]: DEBUG nova.network.neutron [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.538089] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f5acd420-a752-4256-9062-2547ae22223e tempest-ServersTestManualDisk-1131741670 tempest-ServersTestManualDisk-1131741670-project-member] Lock "d46bdbc6-e691-44b7-8e9c-3bedb1275e93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.296s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.672397] env[62558]: DEBUG nova.scheduler.client.report [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.883346] env[62558]: INFO nova.compute.manager [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] [instance: 8e7dcc93-2079-4857-8f18-513958ef4e0a] Took 1.02 seconds to deallocate network for instance. [ 679.043019] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.177934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.178665] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.186087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.214s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.566434] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.688979] env[62558]: DEBUG nova.compute.utils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.691622] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.691782] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.750936] env[62558]: DEBUG nova.policy [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8434f838ad104f9491f9cf4df7093672', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49dd92c04a6b48c5a71b889d5e016135', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.916246] env[62558]: INFO nova.scheduler.client.report [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Deleted allocations for instance 8e7dcc93-2079-4857-8f18-513958ef4e0a [ 680.119275] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Successfully created port: 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.177008] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601cbd7d-8fda-4258-a559-f1ad8f6bd3de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.184640] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd59aee-6c7f-46fe-a8e7-ec643c00e97a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.213196] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.216791] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58e42a5-0670-46eb-ab96-8cd59473b8fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.225206] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261a7499-8879-49f6-b7d9-8a44a97f0d17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.238655] env[62558]: DEBUG nova.compute.provider_tree [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.428569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8392d8f5-5e8e-452b-a668-77d62c2d686a tempest-AttachInterfacesV270Test-674738069 tempest-AttachInterfacesV270Test-674738069-project-member] Lock "8e7dcc93-2079-4857-8f18-513958ef4e0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.108s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.745485] env[62558]: DEBUG nova.scheduler.client.report [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.931520] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.226269] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.230559] env[62558]: DEBUG nova.compute.manager [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Received event network-changed-8a0fdaeb-5e2b-4b0f-a758-28344a4152a5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.230694] env[62558]: DEBUG nova.compute.manager [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Refreshing instance network info cache due to event network-changed-8a0fdaeb-5e2b-4b0f-a758-28344a4152a5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.230863] env[62558]: DEBUG oslo_concurrency.lockutils [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] Acquiring lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.231011] env[62558]: DEBUG oslo_concurrency.lockutils [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] Acquired lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.231458] env[62558]: DEBUG nova.network.neutron [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Refreshing network info cache for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.252028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.070s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.252737] env[62558]: ERROR nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Traceback (most recent call last): [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.driver.spawn(context, instance, image_meta, [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] vm_ref = self.build_virtual_machine(instance, [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.252737] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] for vif in network_info: [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self._sync_wrapper(fn, *args, **kwargs) [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.wait() [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self[:] = self._gt.wait() [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self._exit_event.wait() [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] result = hub.switch() [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.253052] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return self.greenlet.switch() [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] result = function(*args, **kwargs) [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] return func(*args, **kwargs) [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise e [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] nwinfo = self.network_api.allocate_for_instance( [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] created_port_ids = self._update_ports_for_instance( [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] with excutils.save_and_reraise_exception(): [ 681.253380] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] self.force_reraise() [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise self.value [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] updated_port = self._update_port( [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] _ensure_no_port_binding_failure(port) [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] raise exception.PortBindingFailed(port_id=port['id']) [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] nova.exception.PortBindingFailed: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. [ 681.253710] env[62558]: ERROR nova.compute.manager [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] [ 681.253989] env[62558]: DEBUG nova.compute.utils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.258299] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.583s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.258404] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Build of instance 9274ce52-49af-48df-bcd1-269b8241d93f was re-scheduled: Binding failed for port 28e3c1a0-cd03-4e85-83c9-b4d3c9eb5b3e, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.259335] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.259562] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquiring lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.259702] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Acquired lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.259853] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.284637] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T05:00:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='116e3573-53a1-454a-9874-4506f81e19f5',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1436820147',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.285658] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.285825] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.286015] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.286166] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.286313] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.286530] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.286932] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.287181] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.287449] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.287527] env[62558]: DEBUG nova.virt.hardware [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.288465] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f06637-b20d-4e68-9192-2044b797cdcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.298798] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8b907c-51d6-4602-8bd7-0120d401e5fc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.453611] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.486685] env[62558]: ERROR nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 681.486685] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.486685] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.486685] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.486685] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.486685] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.486685] env[62558]: ERROR nova.compute.manager raise self.value [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.486685] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.486685] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.486685] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.487147] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.487147] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.487147] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 681.487147] env[62558]: ERROR nova.compute.manager [ 681.487147] env[62558]: Traceback (most recent call last): [ 681.487147] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.487147] env[62558]: listener.cb(fileno) [ 681.487147] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.487147] env[62558]: result = function(*args, **kwargs) [ 681.487147] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.487147] env[62558]: return func(*args, **kwargs) [ 681.487147] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.487147] env[62558]: raise e [ 681.487147] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.487147] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 681.487147] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.487147] env[62558]: created_port_ids = self._update_ports_for_instance( [ 681.487147] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.487147] env[62558]: with excutils.save_and_reraise_exception(): [ 681.487147] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.487147] env[62558]: self.force_reraise() [ 681.487147] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.487147] env[62558]: raise self.value [ 681.487147] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.487147] env[62558]: updated_port = self._update_port( [ 681.487147] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.487147] env[62558]: _ensure_no_port_binding_failure(port) [ 681.487147] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.487147] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.487871] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 681.487871] env[62558]: Removing descriptor: 22 [ 681.487871] env[62558]: ERROR nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Traceback (most recent call last): [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] yield resources [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.driver.spawn(context, instance, image_meta, [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.487871] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] vm_ref = self.build_virtual_machine(instance, [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] for vif in network_info: [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self._sync_wrapper(fn, *args, **kwargs) [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.wait() [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self[:] = self._gt.wait() [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self._exit_event.wait() [ 681.488264] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] result = hub.switch() [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self.greenlet.switch() [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] result = function(*args, **kwargs) [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return func(*args, **kwargs) [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise e [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] nwinfo = self.network_api.allocate_for_instance( [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.488717] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] created_port_ids = self._update_ports_for_instance( [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] with excutils.save_and_reraise_exception(): [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.force_reraise() [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise self.value [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] updated_port = self._update_port( [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] _ensure_no_port_binding_failure(port) [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.489169] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise exception.PortBindingFailed(port_id=port['id']) [ 681.489601] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 681.489601] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] [ 681.489601] env[62558]: INFO nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Terminating instance [ 681.490247] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.764139] env[62558]: DEBUG nova.network.neutron [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.789090] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.858511] env[62558]: DEBUG nova.network.neutron [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.870163] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.140489] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a4cb69-566d-43aa-bd4d-bd78dfc71cf5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.148403] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a72ba4c-34ea-4a9f-8cd5-7fd839e9e011 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.179855] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d46c7f-3488-49e8-8adf-f173f232ab22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.185289] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1f1074-0ca3-42be-bf04-c2c2e921eef2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.201801] env[62558]: DEBUG nova.compute.provider_tree [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.362604] env[62558]: DEBUG oslo_concurrency.lockutils [req-29ac2735-a97b-4f41-aaa6-ee4a91ef9c14 req-cda2bb80-b141-4475-bad3-2d6773a481ca service nova] Releasing lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.362604] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquired lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.362604] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.376206] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Releasing lock "refresh_cache-9274ce52-49af-48df-bcd1-269b8241d93f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.376206] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.376206] env[62558]: DEBUG nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.376206] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.400205] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.707078] env[62558]: DEBUG nova.scheduler.client.report [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.883851] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.903157] env[62558]: DEBUG nova.network.neutron [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.934580] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.211553] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.957s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.212305] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Traceback (most recent call last): [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.driver.spawn(context, instance, image_meta, [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] vm_ref = self.build_virtual_machine(instance, [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.212305] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] for vif in network_info: [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self._sync_wrapper(fn, *args, **kwargs) [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.wait() [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self[:] = self._gt.wait() [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self._exit_event.wait() [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] result = hub.switch() [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.212662] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return self.greenlet.switch() [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] result = function(*args, **kwargs) [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] return func(*args, **kwargs) [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise e [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] nwinfo = self.network_api.allocate_for_instance( [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] created_port_ids = self._update_ports_for_instance( [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] with excutils.save_and_reraise_exception(): [ 683.213043] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] self.force_reraise() [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise self.value [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] updated_port = self._update_port( [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] _ensure_no_port_binding_failure(port) [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] raise exception.PortBindingFailed(port_id=port['id']) [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] nova.exception.PortBindingFailed: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. [ 683.213478] env[62558]: ERROR nova.compute.manager [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] [ 683.213815] env[62558]: DEBUG nova.compute.utils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.214232] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.726s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.217228] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Build of instance 88cee397-bcbd-428c-b203-e5d9b6acbbb2 was re-scheduled: Binding failed for port dd165904-c6de-4b98-8adf-d2d1308a20aa, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.217657] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.222092] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.222306] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.222485] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.299394] env[62558]: DEBUG nova.compute.manager [req-fa40bf81-18fe-4991-b245-b8baa7764e67 req-b845bb74-e52c-47b6-9ee0-5115c9d61d16 service nova] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Received event network-vif-deleted-8a0fdaeb-5e2b-4b0f-a758-28344a4152a5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.407472] env[62558]: INFO nova.compute.manager [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] [instance: 9274ce52-49af-48df-bcd1-269b8241d93f] Took 1.03 seconds to deallocate network for instance. [ 683.440028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Releasing lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.440028] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.440028] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.440028] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69ac62fb-b917-411f-b12b-130a8c5d3c9e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.449752] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22962f1f-a332-4a44-b3d9-4836e6071bb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.471229] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dca0a819-d11d-437e-9a4c-da4279a9d553 could not be found. [ 683.471442] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.471619] env[62558]: INFO nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.471853] env[62558]: DEBUG oslo.service.loopingcall [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.472061] env[62558]: DEBUG nova.compute.manager [-] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.472155] env[62558]: DEBUG nova.network.neutron [-] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.486843] env[62558]: DEBUG nova.network.neutron [-] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.747527] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.819154] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.991019] env[62558]: DEBUG nova.network.neutron [-] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.119222] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c1bfdb-fd11-4ac7-b5a2-bca832f3f8ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.126926] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e04f82-f66a-4ab1-874b-86e3d9738611 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.166167] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227545c7-08b0-40fb-b651-14de1ec0072a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.173115] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62720019-430d-4bd7-82eb-852db8c1e6e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.186164] env[62558]: DEBUG nova.compute.provider_tree [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.321769] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-88cee397-bcbd-428c-b203-e5d9b6acbbb2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.321769] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.321930] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.322460] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.337490] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.437448] env[62558]: INFO nova.scheduler.client.report [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Deleted allocations for instance 9274ce52-49af-48df-bcd1-269b8241d93f [ 684.491905] env[62558]: INFO nova.compute.manager [-] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Took 1.02 seconds to deallocate network for instance. [ 684.494276] env[62558]: DEBUG nova.compute.claims [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.494442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.689261] env[62558]: DEBUG nova.scheduler.client.report [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.840523] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.945974] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9f2bcc-0413-47fc-ba4b-711d2e647b3b tempest-InstanceActionsNegativeTestJSON-1262331644 tempest-InstanceActionsNegativeTestJSON-1262331644-project-member] Lock "9274ce52-49af-48df-bcd1-269b8241d93f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.794s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.194786] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.195427] env[62558]: ERROR nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] Traceback (most recent call last): [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.driver.spawn(context, instance, image_meta, [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] vm_ref = self.build_virtual_machine(instance, [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.195427] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] for vif in network_info: [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self._sync_wrapper(fn, *args, **kwargs) [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.wait() [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self[:] = self._gt.wait() [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self._exit_event.wait() [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] result = hub.switch() [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.195795] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return self.greenlet.switch() [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] result = function(*args, **kwargs) [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] return func(*args, **kwargs) [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise e [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] nwinfo = self.network_api.allocate_for_instance( [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] created_port_ids = self._update_ports_for_instance( [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] with excutils.save_and_reraise_exception(): [ 685.196360] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] self.force_reraise() [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise self.value [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] updated_port = self._update_port( [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] _ensure_no_port_binding_failure(port) [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] raise exception.PortBindingFailed(port_id=port['id']) [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 685.196744] env[62558]: ERROR nova.compute.manager [instance: ef19262f-77be-470a-b510-29b8016da5a5] [ 685.197651] env[62558]: DEBUG nova.compute.utils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.197802] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Build of instance ef19262f-77be-470a-b510-29b8016da5a5 was re-scheduled: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.198259] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.198564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.198721] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.198886] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.199934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.184s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.201468] env[62558]: INFO nova.compute.claims [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.343309] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: 88cee397-bcbd-428c-b203-e5d9b6acbbb2] Took 1.02 seconds to deallocate network for instance. [ 685.467277] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.730079] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.874088] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.993666] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.381196] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-ef19262f-77be-470a-b510-29b8016da5a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.381445] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.381623] env[62558]: DEBUG nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.384019] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.389665] env[62558]: INFO nova.scheduler.client.report [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleted allocations for instance 88cee397-bcbd-428c-b203-e5d9b6acbbb2 [ 686.408305] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.729251] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1146863-b0b8-4d2c-8b22-ad62522c8d28 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.737076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a006e88c-4a6b-48ca-acb2-b795fd876910 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.769230] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e0a2a1-6920-41ab-ba05-997208066318 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.776793] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97ad2e0-a112-48d3-91ce-898bc20ac8d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.790372] env[62558]: DEBUG nova.compute.provider_tree [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.904720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "88cee397-bcbd-428c-b203-e5d9b6acbbb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.591s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.910995] env[62558]: DEBUG nova.network.neutron [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.293177] env[62558]: DEBUG nova.scheduler.client.report [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.407492] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.413491] env[62558]: INFO nova.compute.manager [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ef19262f-77be-470a-b510-29b8016da5a5] Took 1.03 seconds to deallocate network for instance. [ 687.798986] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.599s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.799549] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.807109] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.097s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.934873] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.307640] env[62558]: DEBUG nova.compute.utils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.309182] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 688.309442] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 688.355567] env[62558]: DEBUG nova.policy [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3c4474feb3a4f5ea478ad5e30d9ea49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '480afedc5a8845e8920815407f4485dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.639517] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.639728] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.661230] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.661577] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.696016] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.696271] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.713757] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Successfully created port: 1672734d-06c0-4100-9b1c-372c9a4c47ec {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.762387] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098718ec-9360-4eb0-a3be-fcb62f36b092 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.770901] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aef25a9-a567-4a6f-a8d9-49270ba5c73d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.799938] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a06b8f-31e2-4c28-9c9f-623ce36949af {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.807542] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720eeee8-f415-423a-8736-21797eb964da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.813255] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.825660] env[62558]: DEBUG nova.compute.provider_tree [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.332437] env[62558]: DEBUG nova.scheduler.client.report [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.454467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2fdc994c-16d7-4309-b9f1-9cb40a380881 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ef19262f-77be-470a-b510-29b8016da5a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.109s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.455111] env[62558]: Traceback (most recent call last): [ 689.455210] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.455210] env[62558]: self.driver.spawn(context, instance, image_meta, [ 689.455210] env[62558]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.455210] env[62558]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.455210] env[62558]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.455210] env[62558]: vm_ref = self.build_virtual_machine(instance, [ 689.455210] env[62558]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.455210] env[62558]: vif_infos = vmwarevif.get_vif_info(self._session, [ 689.455210] env[62558]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.455210] env[62558]: for vif in network_info: [ 689.455210] env[62558]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.455210] env[62558]: return self._sync_wrapper(fn, *args, **kwargs) [ 689.455210] env[62558]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.455210] env[62558]: self.wait() [ 689.455210] env[62558]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.455210] env[62558]: self[:] = self._gt.wait() [ 689.455210] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.455210] env[62558]: return self._exit_event.wait() [ 689.455210] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.455210] env[62558]: result = hub.switch() [ 689.455210] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.455210] env[62558]: return self.greenlet.switch() [ 689.455210] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.455210] env[62558]: result = function(*args, **kwargs) [ 689.455210] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.455210] env[62558]: return func(*args, **kwargs) [ 689.455210] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.455210] env[62558]: raise e [ 689.455210] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.455210] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 689.455210] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.455210] env[62558]: created_port_ids = self._update_ports_for_instance( [ 689.460148] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.460148] env[62558]: with excutils.save_and_reraise_exception(): [ 689.460148] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.460148] env[62558]: self.force_reraise() [ 689.460148] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.460148] env[62558]: raise self.value [ 689.460148] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.460148] env[62558]: updated_port = self._update_port( [ 689.460148] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.460148] env[62558]: _ensure_no_port_binding_failure(port) [ 689.460148] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.460148] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.460148] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 689.460148] env[62558]: During handling of the above exception, another exception occurred: [ 689.460148] env[62558]: Traceback (most recent call last): [ 689.460148] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 689.460148] env[62558]: self._build_and_run_instance(context, instance, image, [ 689.460148] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 689.460148] env[62558]: raise exception.RescheduledException( [ 689.460148] env[62558]: nova.exception.RescheduledException: Build of instance ef19262f-77be-470a-b510-29b8016da5a5 was re-scheduled: Binding failed for port 23f5e7bb-7d33-4809-b60a-9dc4e19a6549, please check neutron logs for more information. [ 689.460148] env[62558]: During handling of the above exception, another exception occurred: [ 689.460148] env[62558]: Traceback (most recent call last): [ 689.460148] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 689.460148] env[62558]: func(*args, **kwargs) [ 689.460148] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.460148] env[62558]: return func(*args, **kwargs) [ 689.460148] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 689.460148] env[62558]: return f(*args, **kwargs) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 689.461087] env[62558]: result = self._do_build_and_run_instance(*args, **kwargs) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 689.461087] env[62558]: with excutils.save_and_reraise_exception(): [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.461087] env[62558]: self.force_reraise() [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.461087] env[62558]: raise self.value [ 689.461087] env[62558]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 689.461087] env[62558]: return f(self, context, *args, **kw) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 689.461087] env[62558]: with excutils.save_and_reraise_exception(): [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.461087] env[62558]: self.force_reraise() [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.461087] env[62558]: raise self.value [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 689.461087] env[62558]: return function(self, context, *args, **kwargs) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 689.461087] env[62558]: return function(self, context, *args, **kwargs) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 689.461087] env[62558]: return function(self, context, *args, **kwargs) [ 689.461087] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 689.461087] env[62558]: instance.save() [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 689.461087] env[62558]: updates, result = self.indirection_api.object_action( [ 689.461087] env[62558]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 689.461087] env[62558]: return cctxt.call(context, 'object_action', objinst=objinst, [ 689.461087] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 689.461087] env[62558]: result = self.transport._send( [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 689.461842] env[62558]: return self._driver.send(target, ctxt, message, [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 689.461842] env[62558]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 689.461842] env[62558]: raise result [ 689.461842] env[62558]: nova.exception_Remote.InstanceNotFound_Remote: Instance ef19262f-77be-470a-b510-29b8016da5a5 could not be found. [ 689.461842] env[62558]: Traceback (most recent call last): [ 689.461842] env[62558]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 689.461842] env[62558]: return getattr(target, method)(*args, **kwargs) [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 689.461842] env[62558]: return fn(self, *args, **kwargs) [ 689.461842] env[62558]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 689.461842] env[62558]: old_ref, inst_ref = db.instance_update_and_get_original( [ 689.461842] env[62558]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 689.461842] env[62558]: return f(*args, **kwargs) [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 689.461842] env[62558]: with excutils.save_and_reraise_exception() as ectxt: [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.461842] env[62558]: self.force_reraise() [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.461842] env[62558]: raise self.value [ 689.461842] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 689.461842] env[62558]: return f(*args, **kwargs) [ 689.461842] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 689.461842] env[62558]: return f(context, *args, **kwargs) [ 689.461842] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 689.461842] env[62558]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 689.461842] env[62558]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 689.462668] env[62558]: raise exception.InstanceNotFound(instance_id=uuid) [ 689.462668] env[62558]: nova.exception.InstanceNotFound: Instance ef19262f-77be-470a-b510-29b8016da5a5 could not be found. [ 689.481711] env[62558]: DEBUG nova.compute.manager [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Received event network-changed-1672734d-06c0-4100-9b1c-372c9a4c47ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.481892] env[62558]: DEBUG nova.compute.manager [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Refreshing instance network info cache due to event network-changed-1672734d-06c0-4100-9b1c-372c9a4c47ec. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 689.482139] env[62558]: DEBUG oslo_concurrency.lockutils [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] Acquiring lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.482292] env[62558]: DEBUG oslo_concurrency.lockutils [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] Acquired lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.482447] env[62558]: DEBUG nova.network.neutron [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Refreshing network info cache for port 1672734d-06c0-4100-9b1c-372c9a4c47ec {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.606199] env[62558]: ERROR nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 689.606199] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.606199] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.606199] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.606199] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.606199] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.606199] env[62558]: ERROR nova.compute.manager raise self.value [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.606199] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.606199] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.606199] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.606774] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.606774] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.606774] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 689.606774] env[62558]: ERROR nova.compute.manager [ 689.606774] env[62558]: Traceback (most recent call last): [ 689.606774] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.606774] env[62558]: listener.cb(fileno) [ 689.606774] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.606774] env[62558]: result = function(*args, **kwargs) [ 689.606774] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.606774] env[62558]: return func(*args, **kwargs) [ 689.606774] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.606774] env[62558]: raise e [ 689.606774] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.606774] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 689.606774] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.606774] env[62558]: created_port_ids = self._update_ports_for_instance( [ 689.606774] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.606774] env[62558]: with excutils.save_and_reraise_exception(): [ 689.606774] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.606774] env[62558]: self.force_reraise() [ 689.606774] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.606774] env[62558]: raise self.value [ 689.606774] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.606774] env[62558]: updated_port = self._update_port( [ 689.606774] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.606774] env[62558]: _ensure_no_port_binding_failure(port) [ 689.606774] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.606774] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.607901] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 689.607901] env[62558]: Removing descriptor: 15 [ 689.837251] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.840550] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.841329] env[62558]: ERROR nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Traceback (most recent call last): [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.driver.spawn(context, instance, image_meta, [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] vm_ref = self.build_virtual_machine(instance, [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.841329] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] for vif in network_info: [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self._sync_wrapper(fn, *args, **kwargs) [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.wait() [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self[:] = self._gt.wait() [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self._exit_event.wait() [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] result = hub.switch() [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.841740] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return self.greenlet.switch() [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] result = function(*args, **kwargs) [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] return func(*args, **kwargs) [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise e [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] nwinfo = self.network_api.allocate_for_instance( [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] created_port_ids = self._update_ports_for_instance( [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] with excutils.save_and_reraise_exception(): [ 689.842176] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] self.force_reraise() [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise self.value [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] updated_port = self._update_port( [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] _ensure_no_port_binding_failure(port) [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] raise exception.PortBindingFailed(port_id=port['id']) [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] nova.exception.PortBindingFailed: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. [ 689.842676] env[62558]: ERROR nova.compute.manager [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] [ 689.843075] env[62558]: DEBUG nova.compute.utils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.843480] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.748s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.844713] env[62558]: INFO nova.compute.claims [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.848035] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Build of instance 2236b710-2141-49a2-b3f5-9ed71090e1db was re-scheduled: Binding failed for port b1ebb42b-0b23-4c39-84c2-ad6339bfeb5d, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 689.848758] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 689.848758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquiring lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.848929] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Acquired lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.849082] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.871353] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.874019] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.874019] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.874019] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.874019] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.874019] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.874249] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.874249] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.874249] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.874249] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.874249] env[62558]: DEBUG nova.virt.hardware [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.874386] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1a55b1-e1cd-4698-b5c0-58b135057588 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.882105] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd73ed2a-f899-4f1e-9427-18f9d4aef60e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.896469] env[62558]: ERROR nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Traceback (most recent call last): [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] yield resources [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.driver.spawn(context, instance, image_meta, [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] vm_ref = self.build_virtual_machine(instance, [ 689.896469] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] for vif in network_info: [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return self._sync_wrapper(fn, *args, **kwargs) [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.wait() [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self[:] = self._gt.wait() [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return self._exit_event.wait() [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.896879] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] current.throw(*self._exc) [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] result = function(*args, **kwargs) [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return func(*args, **kwargs) [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise e [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] nwinfo = self.network_api.allocate_for_instance( [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] created_port_ids = self._update_ports_for_instance( [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] with excutils.save_and_reraise_exception(): [ 689.897263] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.force_reraise() [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise self.value [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] updated_port = self._update_port( [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] _ensure_no_port_binding_failure(port) [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise exception.PortBindingFailed(port_id=port['id']) [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 689.897658] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] [ 689.897658] env[62558]: INFO nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Terminating instance [ 689.899057] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.957968] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.002960] env[62558]: DEBUG nova.network.neutron [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.084000] env[62558]: DEBUG nova.network.neutron [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.368879] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.438388] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.480397] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.586551] env[62558]: DEBUG oslo_concurrency.lockutils [req-1360bfae-496e-43dd-9221-40eef0f8b179 req-5fc2a228-7f26-4287-b1d2-4af723322307 service nova] Releasing lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.587327] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.587327] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.940890] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Releasing lock "refresh_cache-2236b710-2141-49a2-b3f5-9ed71090e1db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.941158] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 690.941319] env[62558]: DEBUG nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.941484] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.957482] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.104903] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.181658] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821af2de-d044-431f-ae90-2674cd7a1145 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.190118] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8684b0-21f4-46a4-9bf9-5f22bb52c907 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.220424] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.222096] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af86f6a6-90ea-4fc9-92c6-34439fc34b85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.230721] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9fe7f5-7d89-4361-b376-dff48c3cfde8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.244417] env[62558]: DEBUG nova.compute.provider_tree [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.462925] env[62558]: DEBUG nova.network.neutron [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.506963] env[62558]: DEBUG nova.compute.manager [req-dbe08452-5224-4b6d-bae9-e94da11ca146 req-cc40ebf0-56fb-42e2-bcaa-becd0cfc836d service nova] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Received event network-vif-deleted-1672734d-06c0-4100-9b1c-372c9a4c47ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.725571] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.726145] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.726345] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.726653] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ceae19c4-417f-42db-8ff6-9f2858801330 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.737289] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7ebdf7-3dee-411a-ae72-04f7196f3d8f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.747677] env[62558]: DEBUG nova.scheduler.client.report [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.762733] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65e8971a-a149-4a11-92ad-7c5c8874530c could not be found. [ 691.762937] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.763128] env[62558]: INFO nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 691.763366] env[62558]: DEBUG oslo.service.loopingcall [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.764109] env[62558]: DEBUG nova.compute.manager [-] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.764205] env[62558]: DEBUG nova.network.neutron [-] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.779390] env[62558]: DEBUG nova.network.neutron [-] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.965797] env[62558]: INFO nova.compute.manager [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] [instance: 2236b710-2141-49a2-b3f5-9ed71090e1db] Took 1.02 seconds to deallocate network for instance. [ 692.252253] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.252798] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.255944] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.156s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.257221] env[62558]: INFO nova.compute.claims [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.281795] env[62558]: DEBUG nova.network.neutron [-] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.757520] env[62558]: DEBUG nova.compute.utils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.759130] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.759398] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 692.783841] env[62558]: INFO nova.compute.manager [-] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Took 1.02 seconds to deallocate network for instance. [ 692.785924] env[62558]: DEBUG nova.compute.claims [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.786108] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.812303] env[62558]: DEBUG nova.policy [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61cb5cd30d434f1eb594dd4521dab2d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e1d3e1dcd64e918c1e436e691e290b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.999562] env[62558]: INFO nova.scheduler.client.report [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Deleted allocations for instance 2236b710-2141-49a2-b3f5-9ed71090e1db [ 693.058634] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Successfully created port: fc0ee510-649e-46bc-a194-f0c19b5cba74 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.263230] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.512180] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b181ccc5-b0b1-4745-b453-bcf40e3567f3 tempest-ServerActionsTestOtherB-1452240497 tempest-ServerActionsTestOtherB-1452240497-project-member] Lock "2236b710-2141-49a2-b3f5-9ed71090e1db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.869s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.648954] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a08f101-f352-42a3-85a4-96d25a2e5e40 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.656790] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f7bcbb-c6fc-4007-b558-c4a64766acaf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.691130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c89f7d-a93c-42e9-9391-b2cf4526ab94 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.701387] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0036b867-9ecb-4072-a8b4-e813f6f473e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.713658] env[62558]: DEBUG nova.compute.provider_tree [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.824244] env[62558]: DEBUG nova.compute.manager [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Received event network-changed-fc0ee510-649e-46bc-a194-f0c19b5cba74 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.824464] env[62558]: DEBUG nova.compute.manager [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Refreshing instance network info cache due to event network-changed-fc0ee510-649e-46bc-a194-f0c19b5cba74. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.824805] env[62558]: DEBUG oslo_concurrency.lockutils [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] Acquiring lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.824805] env[62558]: DEBUG oslo_concurrency.lockutils [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] Acquired lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.824914] env[62558]: DEBUG nova.network.neutron [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Refreshing network info cache for port fc0ee510-649e-46bc-a194-f0c19b5cba74 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.944595] env[62558]: ERROR nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 693.944595] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.944595] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.944595] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.944595] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.944595] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.944595] env[62558]: ERROR nova.compute.manager raise self.value [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.944595] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.944595] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.944595] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.945208] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.945208] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.945208] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 693.945208] env[62558]: ERROR nova.compute.manager [ 693.945208] env[62558]: Traceback (most recent call last): [ 693.945208] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.945208] env[62558]: listener.cb(fileno) [ 693.945208] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.945208] env[62558]: result = function(*args, **kwargs) [ 693.945208] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.945208] env[62558]: return func(*args, **kwargs) [ 693.945208] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.945208] env[62558]: raise e [ 693.945208] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.945208] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 693.945208] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.945208] env[62558]: created_port_ids = self._update_ports_for_instance( [ 693.945208] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.945208] env[62558]: with excutils.save_and_reraise_exception(): [ 693.945208] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.945208] env[62558]: self.force_reraise() [ 693.945208] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.945208] env[62558]: raise self.value [ 693.945208] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.945208] env[62558]: updated_port = self._update_port( [ 693.945208] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.945208] env[62558]: _ensure_no_port_binding_failure(port) [ 693.945208] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.945208] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.946071] env[62558]: nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 693.946071] env[62558]: Removing descriptor: 15 [ 694.014459] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.216931] env[62558]: DEBUG nova.scheduler.client.report [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.277478] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.301469] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.301710] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.301864] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.302046] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.302190] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.302328] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.302530] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.302683] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.302843] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.302999] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.303184] env[62558]: DEBUG nova.virt.hardware [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.304286] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea88d020-c6e1-4198-a773-a4ee7b2a9b46 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.313446] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31366139-5787-4949-9642-f865f87f018d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.329851] env[62558]: ERROR nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] Traceback (most recent call last): [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] yield resources [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.driver.spawn(context, instance, image_meta, [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] vm_ref = self.build_virtual_machine(instance, [ 694.329851] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] for vif in network_info: [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return self._sync_wrapper(fn, *args, **kwargs) [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.wait() [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self[:] = self._gt.wait() [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return self._exit_event.wait() [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.330240] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] current.throw(*self._exc) [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] result = function(*args, **kwargs) [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return func(*args, **kwargs) [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise e [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] nwinfo = self.network_api.allocate_for_instance( [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] created_port_ids = self._update_ports_for_instance( [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] with excutils.save_and_reraise_exception(): [ 694.330568] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.force_reraise() [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise self.value [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] updated_port = self._update_port( [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] _ensure_no_port_binding_failure(port) [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise exception.PortBindingFailed(port_id=port['id']) [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 694.330894] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] [ 694.330894] env[62558]: INFO nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Terminating instance [ 694.332089] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.348586] env[62558]: DEBUG nova.network.neutron [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.487048] env[62558]: DEBUG nova.network.neutron [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.546759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.729858] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.730427] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.737142] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.171s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.738743] env[62558]: INFO nova.compute.claims [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.991603] env[62558]: DEBUG oslo_concurrency.lockutils [req-85bc7ae7-b53b-4011-853c-872f61ea206f req-ab6ec12b-1a6a-4152-8be7-0229786116a5 service nova] Releasing lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.991603] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.991603] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.246840] env[62558]: DEBUG nova.compute.utils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.253219] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.253499] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 695.323058] env[62558]: DEBUG nova.policy [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '347eb0cd1ed6466088feac8e0462f542', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8dd44300fa874f858ef35df6e911d566', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.515322] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.607612] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.740916] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Successfully created port: 725c06cc-3cf6-4685-bfd8-40182e6b717e {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.753524] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.881919] env[62558]: DEBUG nova.compute.manager [req-f48d8149-e2c0-4e55-9d8a-869f22eb2ab7 req-87473017-67bd-4ea0-9669-034d1ac1ca89 service nova] [instance: b482dcca-ae32-425d-a097-f9af69785835] Received event network-vif-deleted-fc0ee510-649e-46bc-a194-f0c19b5cba74 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.109836] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.110271] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.110458] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.110825] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3be672bd-add1-4aa4-aef9-dfaa90c662a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.126278] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8256762-7c16-4ac4-80b3-f573490fd8cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.153931] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b482dcca-ae32-425d-a097-f9af69785835 could not be found. [ 696.154193] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.154376] env[62558]: INFO nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Took 0.04 seconds to destroy the instance on the hypervisor. [ 696.154619] env[62558]: DEBUG oslo.service.loopingcall [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.156972] env[62558]: DEBUG nova.compute.manager [-] [instance: b482dcca-ae32-425d-a097-f9af69785835] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.157082] env[62558]: DEBUG nova.network.neutron [-] [instance: b482dcca-ae32-425d-a097-f9af69785835] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.185613] env[62558]: DEBUG nova.network.neutron [-] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.200956] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20081c8f-50e5-4968-a067-3141c8c2a8b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.209407] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67767de-9307-4cd8-8749-8f382db5a66d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.241237] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390d8c57-b6d5-4122-a0ec-6a55308ccc85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.248783] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb74dc1-149a-4073-a263-94ea1918c12e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.265424] env[62558]: DEBUG nova.compute.provider_tree [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.690804] env[62558]: DEBUG nova.network.neutron [-] [instance: b482dcca-ae32-425d-a097-f9af69785835] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.767862] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 696.770523] env[62558]: DEBUG nova.scheduler.client.report [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.796140] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.796419] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.796573] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.796751] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.796893] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.797050] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.797264] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.797425] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.797587] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.797747] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.797914] env[62558]: DEBUG nova.virt.hardware [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.798806] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc038a0-2132-4725-871e-ceef898158c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.807937] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf42f1f-b697-4584-b0ae-8f01230b4eac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.838439] env[62558]: ERROR nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 696.838439] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.838439] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.838439] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.838439] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.838439] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.838439] env[62558]: ERROR nova.compute.manager raise self.value [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.838439] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 696.838439] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.838439] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 696.839389] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.839389] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 696.839389] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 696.839389] env[62558]: ERROR nova.compute.manager [ 696.839389] env[62558]: Traceback (most recent call last): [ 696.839389] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 696.839389] env[62558]: listener.cb(fileno) [ 696.839389] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.839389] env[62558]: result = function(*args, **kwargs) [ 696.839389] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.839389] env[62558]: return func(*args, **kwargs) [ 696.839389] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.839389] env[62558]: raise e [ 696.839389] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.839389] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 696.839389] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.839389] env[62558]: created_port_ids = self._update_ports_for_instance( [ 696.839389] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.839389] env[62558]: with excutils.save_and_reraise_exception(): [ 696.839389] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.839389] env[62558]: self.force_reraise() [ 696.839389] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.839389] env[62558]: raise self.value [ 696.839389] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.839389] env[62558]: updated_port = self._update_port( [ 696.839389] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.839389] env[62558]: _ensure_no_port_binding_failure(port) [ 696.839389] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.839389] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 696.840723] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 696.840723] env[62558]: Removing descriptor: 15 [ 696.840723] env[62558]: ERROR nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] Traceback (most recent call last): [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] yield resources [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.driver.spawn(context, instance, image_meta, [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.840723] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] vm_ref = self.build_virtual_machine(instance, [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] for vif in network_info: [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self._sync_wrapper(fn, *args, **kwargs) [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.wait() [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self[:] = self._gt.wait() [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self._exit_event.wait() [ 696.841284] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] result = hub.switch() [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self.greenlet.switch() [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] result = function(*args, **kwargs) [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return func(*args, **kwargs) [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise e [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] nwinfo = self.network_api.allocate_for_instance( [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.841830] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] created_port_ids = self._update_ports_for_instance( [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] with excutils.save_and_reraise_exception(): [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.force_reraise() [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise self.value [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] updated_port = self._update_port( [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] _ensure_no_port_binding_failure(port) [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.842426] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise exception.PortBindingFailed(port_id=port['id']) [ 696.842945] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 696.842945] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] [ 696.842945] env[62558]: INFO nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Terminating instance [ 696.842945] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquiring lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.842945] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquired lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.842945] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.193087] env[62558]: INFO nova.compute.manager [-] [instance: b482dcca-ae32-425d-a097-f9af69785835] Took 1.04 seconds to deallocate network for instance. [ 697.195777] env[62558]: DEBUG nova.compute.claims [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.195983] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.275148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.275709] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.279200] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.825s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.279883] env[62558]: INFO nova.compute.claims [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.359872] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.450853] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.784299] env[62558]: DEBUG nova.compute.utils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.787942] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.788462] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.841814] env[62558]: DEBUG nova.policy [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4bf855164714b32b792e97c1da99b8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0d6a928119744f3a933e6ecc49b1686', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.909541] env[62558]: DEBUG nova.compute.manager [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Received event network-changed-725c06cc-3cf6-4685-bfd8-40182e6b717e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.909740] env[62558]: DEBUG nova.compute.manager [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Refreshing instance network info cache due to event network-changed-725c06cc-3cf6-4685-bfd8-40182e6b717e. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 697.909925] env[62558]: DEBUG oslo_concurrency.lockutils [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] Acquiring lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.955471] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Releasing lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.955954] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.956167] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.956504] env[62558]: DEBUG oslo_concurrency.lockutils [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] Acquired lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.956676] env[62558]: DEBUG nova.network.neutron [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Refreshing network info cache for port 725c06cc-3cf6-4685-bfd8-40182e6b717e {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 697.957773] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92a05093-a37a-41f6-802b-b53eae58f772 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.968091] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f89ece9-a3f1-49b1-a755-960a26d42b89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.990491] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8db751a-c495-467c-ae32-addb4387346d could not be found. [ 697.990671] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 697.990851] env[62558]: INFO nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.991104] env[62558]: DEBUG oslo.service.loopingcall [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.991344] env[62558]: DEBUG nova.compute.manager [-] [instance: f8db751a-c495-467c-ae32-addb4387346d] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.991465] env[62558]: DEBUG nova.network.neutron [-] [instance: f8db751a-c495-467c-ae32-addb4387346d] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.008182] env[62558]: DEBUG nova.network.neutron [-] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.133917] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Successfully created port: 52f07dbe-eb8d-4bd8-948e-874e5d745201 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.289941] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.499298] env[62558]: DEBUG nova.network.neutron [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.511401] env[62558]: DEBUG nova.network.neutron [-] [instance: f8db751a-c495-467c-ae32-addb4387346d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.596462] env[62558]: DEBUG nova.network.neutron [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.690907] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a06057-20ea-48d1-9f08-53048e2d39df {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.698772] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0c3729-7dce-4697-bac4-195248db1d9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.729227] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329567d3-142f-4260-b8d0-f8c2b197520e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.736483] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769da96c-9d37-4b94-8e34-9aa82022a67d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.749665] env[62558]: DEBUG nova.compute.provider_tree [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.990451] env[62558]: ERROR nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 698.990451] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.990451] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.990451] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.990451] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.990451] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.990451] env[62558]: ERROR nova.compute.manager raise self.value [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.990451] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.990451] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.990451] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.991095] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.991095] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.991095] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 698.991095] env[62558]: ERROR nova.compute.manager [ 698.991095] env[62558]: Traceback (most recent call last): [ 698.991095] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.991095] env[62558]: listener.cb(fileno) [ 698.991095] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.991095] env[62558]: result = function(*args, **kwargs) [ 698.991095] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.991095] env[62558]: return func(*args, **kwargs) [ 698.991095] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.991095] env[62558]: raise e [ 698.991095] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.991095] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 698.991095] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.991095] env[62558]: created_port_ids = self._update_ports_for_instance( [ 698.991095] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.991095] env[62558]: with excutils.save_and_reraise_exception(): [ 698.991095] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.991095] env[62558]: self.force_reraise() [ 698.991095] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.991095] env[62558]: raise self.value [ 698.991095] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.991095] env[62558]: updated_port = self._update_port( [ 698.991095] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.991095] env[62558]: _ensure_no_port_binding_failure(port) [ 698.991095] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.991095] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.991916] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 698.991916] env[62558]: Removing descriptor: 15 [ 699.014548] env[62558]: INFO nova.compute.manager [-] [instance: f8db751a-c495-467c-ae32-addb4387346d] Took 1.02 seconds to deallocate network for instance. [ 699.016707] env[62558]: DEBUG nova.compute.claims [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.016903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.099911] env[62558]: DEBUG oslo_concurrency.lockutils [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] Releasing lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.099911] env[62558]: DEBUG nova.compute.manager [req-47cdba2b-0d01-477f-a3d1-bfe59e190e00 req-caec799b-1d64-41d8-80c9-02e83e390159 service nova] [instance: f8db751a-c495-467c-ae32-addb4387346d] Received event network-vif-deleted-725c06cc-3cf6-4685-bfd8-40182e6b717e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.253075] env[62558]: DEBUG nova.scheduler.client.report [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.303459] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.329294] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.329294] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.329294] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.329512] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.329968] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.329968] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.329968] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.330172] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.330316] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.330474] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.330639] env[62558]: DEBUG nova.virt.hardware [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.331522] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2bfb27-d299-4010-85a7-531482315979 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.339873] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77969d47-5ee0-4f7b-9f24-556e75804fa9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.353339] env[62558]: ERROR nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Traceback (most recent call last): [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] yield resources [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.driver.spawn(context, instance, image_meta, [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] vm_ref = self.build_virtual_machine(instance, [ 699.353339] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] for vif in network_info: [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return self._sync_wrapper(fn, *args, **kwargs) [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.wait() [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self[:] = self._gt.wait() [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return self._exit_event.wait() [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.353689] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] current.throw(*self._exc) [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] result = function(*args, **kwargs) [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return func(*args, **kwargs) [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise e [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] nwinfo = self.network_api.allocate_for_instance( [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] created_port_ids = self._update_ports_for_instance( [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] with excutils.save_and_reraise_exception(): [ 699.354073] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.force_reraise() [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise self.value [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] updated_port = self._update_port( [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] _ensure_no_port_binding_failure(port) [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise exception.PortBindingFailed(port_id=port['id']) [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 699.354464] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] [ 699.354464] env[62558]: INFO nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Terminating instance [ 699.355645] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquiring lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.355802] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquired lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.355960] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.758512] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.759010] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.762559] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.268s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.873148] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.934099] env[62558]: DEBUG nova.compute.manager [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Received event network-changed-52f07dbe-eb8d-4bd8-948e-874e5d745201 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.934311] env[62558]: DEBUG nova.compute.manager [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Refreshing instance network info cache due to event network-changed-52f07dbe-eb8d-4bd8-948e-874e5d745201. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 699.934505] env[62558]: DEBUG oslo_concurrency.lockutils [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] Acquiring lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.963937] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.264256] env[62558]: DEBUG nova.compute.utils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.265660] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.265835] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.310043] env[62558]: DEBUG nova.policy [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.468527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Releasing lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.468901] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.469105] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.469686] env[62558]: DEBUG oslo_concurrency.lockutils [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] Acquired lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.469829] env[62558]: DEBUG nova.network.neutron [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Refreshing network info cache for port 52f07dbe-eb8d-4bd8-948e-874e5d745201 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 700.470830] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-975aa2db-30f0-490d-a64a-366fa33bd98d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.480086] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bcb2e6-8092-4063-8935-972ccf5fa065 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.505762] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b49513e-4c01-4664-be2c-6058a357b0db could not be found. [ 700.505976] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.506173] env[62558]: INFO nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.506524] env[62558]: DEBUG oslo.service.loopingcall [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.510238] env[62558]: DEBUG nova.compute.manager [-] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.510238] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.525556] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.574718] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Successfully created port: a56124a9-37f0-4760-9955-84b74f5385df {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.639193] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb55d89e-4525-4d03-a7cb-14f9fa323508 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.646676] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72e47fe-8691-467e-96f7-f2e8e464ec8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.677647] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc18738-eeb2-493d-ab5f-42c70ca209b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.684421] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfde121-4a38-4269-a947-51860955ee66 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.697636] env[62558]: DEBUG nova.compute.provider_tree [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.771911] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.990477] env[62558]: DEBUG nova.network.neutron [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.028131] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.076208] env[62558]: DEBUG nova.network.neutron [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.201014] env[62558]: DEBUG nova.scheduler.client.report [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.387691] env[62558]: ERROR nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 701.387691] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.387691] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.387691] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.387691] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.387691] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.387691] env[62558]: ERROR nova.compute.manager raise self.value [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.387691] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.387691] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.387691] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.388360] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.388360] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.388360] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 701.388360] env[62558]: ERROR nova.compute.manager [ 701.388360] env[62558]: Traceback (most recent call last): [ 701.388360] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.388360] env[62558]: listener.cb(fileno) [ 701.388360] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.388360] env[62558]: result = function(*args, **kwargs) [ 701.388360] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.388360] env[62558]: return func(*args, **kwargs) [ 701.388360] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.388360] env[62558]: raise e [ 701.388360] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.388360] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 701.388360] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.388360] env[62558]: created_port_ids = self._update_ports_for_instance( [ 701.388360] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.388360] env[62558]: with excutils.save_and_reraise_exception(): [ 701.388360] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.388360] env[62558]: self.force_reraise() [ 701.388360] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.388360] env[62558]: raise self.value [ 701.388360] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.388360] env[62558]: updated_port = self._update_port( [ 701.388360] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.388360] env[62558]: _ensure_no_port_binding_failure(port) [ 701.388360] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.388360] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.389284] env[62558]: nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 701.389284] env[62558]: Removing descriptor: 15 [ 701.531044] env[62558]: INFO nova.compute.manager [-] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Took 1.02 seconds to deallocate network for instance. [ 701.533173] env[62558]: DEBUG nova.compute.claims [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.533344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.578600] env[62558]: DEBUG oslo_concurrency.lockutils [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] Releasing lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.578892] env[62558]: DEBUG nova.compute.manager [req-d5dcbd7b-958a-482d-94d5-a4bf9a6b60f6 req-9c92401a-4ab0-4397-af2d-53d3da563432 service nova] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Received event network-vif-deleted-52f07dbe-eb8d-4bd8-948e-874e5d745201 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.706976] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.707613] env[62558]: ERROR nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Traceback (most recent call last): [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.driver.spawn(context, instance, image_meta, [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] vm_ref = self.build_virtual_machine(instance, [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.707613] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] for vif in network_info: [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self._sync_wrapper(fn, *args, **kwargs) [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.wait() [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self[:] = self._gt.wait() [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self._exit_event.wait() [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] result = hub.switch() [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.707920] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return self.greenlet.switch() [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] result = function(*args, **kwargs) [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] return func(*args, **kwargs) [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise e [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] nwinfo = self.network_api.allocate_for_instance( [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] created_port_ids = self._update_ports_for_instance( [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] with excutils.save_and_reraise_exception(): [ 701.708256] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] self.force_reraise() [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise self.value [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] updated_port = self._update_port( [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] _ensure_no_port_binding_failure(port) [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] raise exception.PortBindingFailed(port_id=port['id']) [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] nova.exception.PortBindingFailed: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. [ 701.708615] env[62558]: ERROR nova.compute.manager [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] [ 701.708948] env[62558]: DEBUG nova.compute.utils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.709572] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.716s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.710995] env[62558]: INFO nova.compute.claims [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.713528] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Build of instance dca0a819-d11d-437e-9a4c-da4279a9d553 was re-scheduled: Binding failed for port 8a0fdaeb-5e2b-4b0f-a758-28344a4152a5, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.713952] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.714191] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquiring lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.715405] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Acquired lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.715405] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.788255] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.812193] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.812443] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.812600] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.812773] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.812929] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.813062] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.814080] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.814080] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.814080] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.814080] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.814080] env[62558]: DEBUG nova.virt.hardware [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.815105] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fd3487-aa59-41f2-a873-3e30718638fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.823315] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aee15dd-4faa-4f4d-9190-f42cbdce9398 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.837287] env[62558]: ERROR nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Traceback (most recent call last): [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] yield resources [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.driver.spawn(context, instance, image_meta, [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] vm_ref = self.build_virtual_machine(instance, [ 701.837287] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] for vif in network_info: [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return self._sync_wrapper(fn, *args, **kwargs) [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.wait() [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self[:] = self._gt.wait() [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return self._exit_event.wait() [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.837726] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] current.throw(*self._exc) [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] result = function(*args, **kwargs) [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return func(*args, **kwargs) [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise e [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] nwinfo = self.network_api.allocate_for_instance( [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] created_port_ids = self._update_ports_for_instance( [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] with excutils.save_and_reraise_exception(): [ 701.838184] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.force_reraise() [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise self.value [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] updated_port = self._update_port( [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] _ensure_no_port_binding_failure(port) [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise exception.PortBindingFailed(port_id=port['id']) [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 701.838603] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] [ 701.838603] env[62558]: INFO nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Terminating instance [ 701.839814] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.839977] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.840160] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.959351] env[62558]: DEBUG nova.compute.manager [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Received event network-changed-a56124a9-37f0-4760-9955-84b74f5385df {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.959553] env[62558]: DEBUG nova.compute.manager [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Refreshing instance network info cache due to event network-changed-a56124a9-37f0-4760-9955-84b74f5385df. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.959742] env[62558]: DEBUG oslo_concurrency.lockutils [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] Acquiring lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.240975] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.306407] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.357340] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.424967] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.809285] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Releasing lock "refresh_cache-dca0a819-d11d-437e-9a4c-da4279a9d553" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.809516] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.809679] env[62558]: DEBUG nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.809863] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.825142] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.927841] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.927981] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 702.928193] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.928531] env[62558]: DEBUG oslo_concurrency.lockutils [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] Acquired lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.928728] env[62558]: DEBUG nova.network.neutron [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Refreshing network info cache for port a56124a9-37f0-4760-9955-84b74f5385df {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 702.929715] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebb49e83-8a5a-401d-83bc-bddc4f06ffd9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.940059] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de131053-cdf2-4b08-98ae-e5646acebdf5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.965548] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4 could not be found. [ 702.965548] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.965548] env[62558]: INFO nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 702.965771] env[62558]: DEBUG oslo.service.loopingcall [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.968246] env[62558]: DEBUG nova.compute.manager [-] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.968361] env[62558]: DEBUG nova.network.neutron [-] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.983395] env[62558]: DEBUG nova.network.neutron [-] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.080023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89371c65-56a6-45ff-8bdf-f319f17a3d45 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.087130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b631e9-d1c4-479e-9e92-5acd08197e72 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.116343] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871a3f95-f1a4-4ec0-82e7-8086b5bb031a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.122963] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fe437e-95f9-4ced-884b-1cf394a822e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.137028] env[62558]: DEBUG nova.compute.provider_tree [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.330564] env[62558]: DEBUG nova.network.neutron [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.447464] env[62558]: DEBUG nova.network.neutron [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.487217] env[62558]: DEBUG nova.network.neutron [-] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.513219] env[62558]: DEBUG nova.network.neutron [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.639184] env[62558]: DEBUG nova.scheduler.client.report [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.833740] env[62558]: INFO nova.compute.manager [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] [instance: dca0a819-d11d-437e-9a4c-da4279a9d553] Took 1.02 seconds to deallocate network for instance. [ 703.990112] env[62558]: INFO nova.compute.manager [-] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Took 1.02 seconds to deallocate network for instance. [ 703.992393] env[62558]: DEBUG nova.compute.claims [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 703.992570] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.015983] env[62558]: DEBUG oslo_concurrency.lockutils [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] Releasing lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.015983] env[62558]: DEBUG nova.compute.manager [req-1a1a9256-006a-4f2e-aac3-78c35801b59c req-ff8ba9df-09a4-4669-b62b-ccdab16e77b9 service nova] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Received event network-vif-deleted-a56124a9-37f0-4760-9955-84b74f5385df {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.144586] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.144746] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.147198] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.212s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.148642] env[62558]: INFO nova.compute.claims [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.653427] env[62558]: DEBUG nova.compute.utils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.656770] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.657174] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.696201] env[62558]: DEBUG nova.policy [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.879368] env[62558]: INFO nova.scheduler.client.report [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Deleted allocations for instance dca0a819-d11d-437e-9a4c-da4279a9d553 [ 704.986538] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Successfully created port: 892b4af7-fadf-4cfb-817a-3e6afbf89fab {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.157703] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.389214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-feb1bacb-0746-446c-af77-61a3b09ce39a tempest-MigrationsAdminTest-756103562 tempest-MigrationsAdminTest-756103562-project-member] Lock "dca0a819-d11d-437e-9a4c-da4279a9d553" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.015s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.487245] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff310b29-e717-4b6a-b247-8b5a0da998a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.495364] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7431ea3c-3e65-48ba-8595-5889fadfda2e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.527674] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb84532-f347-4bcd-b2f4-5a6bc68bbea0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.535151] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c418d7-0a3f-4862-b717-3d22fc931526 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.548123] env[62558]: DEBUG nova.compute.provider_tree [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.642152] env[62558]: DEBUG nova.compute.manager [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Received event network-changed-892b4af7-fadf-4cfb-817a-3e6afbf89fab {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.642152] env[62558]: DEBUG nova.compute.manager [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Refreshing instance network info cache due to event network-changed-892b4af7-fadf-4cfb-817a-3e6afbf89fab. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.642152] env[62558]: DEBUG oslo_concurrency.lockutils [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] Acquiring lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.642152] env[62558]: DEBUG oslo_concurrency.lockutils [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] Acquired lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.642152] env[62558]: DEBUG nova.network.neutron [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Refreshing network info cache for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.821419] env[62558]: ERROR nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 705.821419] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.821419] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.821419] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.821419] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.821419] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.821419] env[62558]: ERROR nova.compute.manager raise self.value [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.821419] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.821419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.821419] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.822197] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.822197] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.822197] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 705.822197] env[62558]: ERROR nova.compute.manager [ 705.822197] env[62558]: Traceback (most recent call last): [ 705.822197] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.822197] env[62558]: listener.cb(fileno) [ 705.822197] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.822197] env[62558]: result = function(*args, **kwargs) [ 705.822197] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.822197] env[62558]: return func(*args, **kwargs) [ 705.822197] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.822197] env[62558]: raise e [ 705.822197] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.822197] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 705.822197] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.822197] env[62558]: created_port_ids = self._update_ports_for_instance( [ 705.822197] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.822197] env[62558]: with excutils.save_and_reraise_exception(): [ 705.822197] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.822197] env[62558]: self.force_reraise() [ 705.822197] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.822197] env[62558]: raise self.value [ 705.822197] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.822197] env[62558]: updated_port = self._update_port( [ 705.822197] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.822197] env[62558]: _ensure_no_port_binding_failure(port) [ 705.822197] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.822197] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.822918] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 705.822918] env[62558]: Removing descriptor: 15 [ 705.894879] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.051080] env[62558]: DEBUG nova.scheduler.client.report [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.165878] env[62558]: DEBUG nova.network.neutron [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.168441] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.209654] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.209925] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.210155] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.210326] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.210473] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.210631] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.210840] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.210998] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.211187] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.211680] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.211680] env[62558]: DEBUG nova.virt.hardware [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.212701] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfd5c07-85a6-4845-aeb7-7ad33a4f02ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.221741] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b167e3-b226-4f8b-a410-358b13355906 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.236609] env[62558]: ERROR nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Traceback (most recent call last): [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] yield resources [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.driver.spawn(context, instance, image_meta, [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] vm_ref = self.build_virtual_machine(instance, [ 706.236609] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] for vif in network_info: [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return self._sync_wrapper(fn, *args, **kwargs) [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.wait() [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self[:] = self._gt.wait() [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return self._exit_event.wait() [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 706.238752] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] current.throw(*self._exc) [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] result = function(*args, **kwargs) [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return func(*args, **kwargs) [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise e [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] nwinfo = self.network_api.allocate_for_instance( [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] created_port_ids = self._update_ports_for_instance( [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] with excutils.save_and_reraise_exception(): [ 706.239150] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.force_reraise() [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise self.value [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] updated_port = self._update_port( [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] _ensure_no_port_binding_failure(port) [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise exception.PortBindingFailed(port_id=port['id']) [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 706.239493] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] [ 706.239493] env[62558]: INFO nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Terminating instance [ 706.240431] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.291019] env[62558]: DEBUG nova.network.neutron [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.418103] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.555781] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.556243] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.560644] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.079s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.562151] env[62558]: INFO nova.compute.claims [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.793564] env[62558]: DEBUG oslo_concurrency.lockutils [req-b6574e56-f0ff-4704-a087-f66b716efb48 req-be47ad62-e20e-40a9-91cd-9e2e51f9cc8e service nova] Releasing lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.793981] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.794179] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.067307] env[62558]: DEBUG nova.compute.utils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.070851] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.071026] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.128034] env[62558]: DEBUG nova.policy [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b578f403d1f442a28f0a56cdf52d0f97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '637a401eb52b439ba0625a7c0a1556bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.319928] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.413055] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Successfully created port: 5e28565a-9cb3-4032-97e1-28a94776d683 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.414267] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.577105] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.671936] env[62558]: DEBUG nova.compute.manager [req-01b1ba3c-0c16-4c4f-a527-7acf424b2139 req-a20cb317-1729-40a6-be4a-bd59b927a19a service nova] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Received event network-vif-deleted-892b4af7-fadf-4cfb-817a-3e6afbf89fab {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.916892] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.917728] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.920607] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.920607] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2c52cd6-7fa1-4248-98fd-103b8c7fed90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.932454] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3d4e31-db8b-4eca-ac2d-666e67ef7d82 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.958885] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2888fde6-8a1b-41ce-a482-cb99acd74a6e could not be found. [ 707.959254] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.959412] env[62558]: INFO nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 707.959752] env[62558]: DEBUG oslo.service.loopingcall [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.962613] env[62558]: DEBUG nova.compute.manager [-] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.962733] env[62558]: DEBUG nova.network.neutron [-] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.001480] env[62558]: DEBUG nova.network.neutron [-] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.027584] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ac02da-849b-4b5e-9d94-f4bf22e3201b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.036867] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd355e9-33e5-4973-843f-f226a31f02e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.070279] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce604a2b-1719-44da-97b7-c1aa73ee273a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.078165] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ab7db6-df76-4acc-a9aa-5b0cfb52d4f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.083262] env[62558]: INFO nova.virt.block_device [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Booting with volume 037f5ce4-4bab-4515-8d66-b684c30581c7 at /dev/sda [ 708.095344] env[62558]: DEBUG nova.compute.provider_tree [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.110631] env[62558]: DEBUG nova.scheduler.client.report [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.144054] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5be6d879-84f4-4d1e-9108-314b92b488bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.152588] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f7974e-dc78-4a24-8024-a3e04d160ae9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.174704] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48e29c35-0041-4a33-a125-69d996b59dd1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.183089] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2daa40de-ba29-457e-a2ac-54a2c1c11b40 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.208349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5cf5e3-f01b-4ac5-8259-0d31255d2992 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.212860] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff20726-f8ed-407c-9c31-eb2badb2a56e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.228199] env[62558]: DEBUG nova.virt.block_device [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating existing volume attachment record: 80735d77-ac9d-4a8e-81a2-fced96242c20 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 708.459334] env[62558]: ERROR nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 708.459334] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.459334] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.459334] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.459334] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.459334] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.459334] env[62558]: ERROR nova.compute.manager raise self.value [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.459334] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.459334] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.459334] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.459869] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.459869] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.459869] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 708.459869] env[62558]: ERROR nova.compute.manager [ 708.459869] env[62558]: Traceback (most recent call last): [ 708.459869] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.459869] env[62558]: listener.cb(fileno) [ 708.459869] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.459869] env[62558]: result = function(*args, **kwargs) [ 708.459869] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.459869] env[62558]: return func(*args, **kwargs) [ 708.459869] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.459869] env[62558]: raise e [ 708.459869] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.459869] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 708.459869] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.459869] env[62558]: created_port_ids = self._update_ports_for_instance( [ 708.459869] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.459869] env[62558]: with excutils.save_and_reraise_exception(): [ 708.459869] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.459869] env[62558]: self.force_reraise() [ 708.459869] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.459869] env[62558]: raise self.value [ 708.459869] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.459869] env[62558]: updated_port = self._update_port( [ 708.459869] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.459869] env[62558]: _ensure_no_port_binding_failure(port) [ 708.459869] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.459869] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.460799] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 708.460799] env[62558]: Removing descriptor: 15 [ 708.504298] env[62558]: DEBUG nova.network.neutron [-] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.616042] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.057s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.616585] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.620556] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.834s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.009220] env[62558]: INFO nova.compute.manager [-] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Took 1.04 seconds to deallocate network for instance. [ 709.010160] env[62558]: DEBUG nova.compute.claims [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.010546] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.124792] env[62558]: DEBUG nova.compute.utils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.124792] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.124792] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.195928] env[62558]: DEBUG nova.policy [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c1ba36c5f0648ea940553d6d7b0e5b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38a5cd0418f841c38e45284f2b12ad53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.524259] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Successfully created port: 7df067c2-30b7-4010-8e39-950390e1cf65 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.579258] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ead03a-69a5-448d-a41a-7597cd797d98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.588447] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b140c90-6f2a-4d62-9017-1de5acff90b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.622871] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1231d5fa-8638-4527-a716-9c94ff31058e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.631190] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc8a383-5936-4d1a-beef-b694d3b72327 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.636864] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.650623] env[62558]: DEBUG nova.compute.provider_tree [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.723268] env[62558]: DEBUG nova.compute.manager [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Received event network-changed-5e28565a-9cb3-4032-97e1-28a94776d683 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.723470] env[62558]: DEBUG nova.compute.manager [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Refreshing instance network info cache due to event network-changed-5e28565a-9cb3-4032-97e1-28a94776d683. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 709.723683] env[62558]: DEBUG oslo_concurrency.lockutils [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] Acquiring lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.723823] env[62558]: DEBUG oslo_concurrency.lockutils [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] Acquired lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.723979] env[62558]: DEBUG nova.network.neutron [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Refreshing network info cache for port 5e28565a-9cb3-4032-97e1-28a94776d683 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.156019] env[62558]: DEBUG nova.scheduler.client.report [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.261484] env[62558]: DEBUG nova.network.neutron [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.360318] env[62558]: DEBUG nova.network.neutron [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.369370] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.369370] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.369370] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.369370] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.369628] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.369628] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.369628] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.369628] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.369628] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.369779] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.370406] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.370740] env[62558]: DEBUG nova.virt.hardware [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.372217] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a52726-0fa2-4577-893a-369d1589fe05 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.386024] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1d7283-cb1e-41a0-a698-c8c308977740 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.400538] env[62558]: ERROR nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Traceback (most recent call last): [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] yield resources [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.driver.spawn(context, instance, image_meta, [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] vm_ref = self.build_virtual_machine(instance, [ 710.400538] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] for vif in network_info: [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return self._sync_wrapper(fn, *args, **kwargs) [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.wait() [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self[:] = self._gt.wait() [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return self._exit_event.wait() [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.400897] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] current.throw(*self._exc) [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] result = function(*args, **kwargs) [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return func(*args, **kwargs) [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise e [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] nwinfo = self.network_api.allocate_for_instance( [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] created_port_ids = self._update_ports_for_instance( [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] with excutils.save_and_reraise_exception(): [ 710.401285] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.force_reraise() [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise self.value [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] updated_port = self._update_port( [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] _ensure_no_port_binding_failure(port) [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise exception.PortBindingFailed(port_id=port['id']) [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 710.401647] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] [ 710.401647] env[62558]: INFO nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Terminating instance [ 710.404348] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquiring lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.605580] env[62558]: ERROR nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 710.605580] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.605580] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.605580] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.605580] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.605580] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.605580] env[62558]: ERROR nova.compute.manager raise self.value [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.605580] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.605580] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.605580] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.606419] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.606419] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.606419] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 710.606419] env[62558]: ERROR nova.compute.manager [ 710.606419] env[62558]: Traceback (most recent call last): [ 710.606419] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.606419] env[62558]: listener.cb(fileno) [ 710.606419] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.606419] env[62558]: result = function(*args, **kwargs) [ 710.606419] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.606419] env[62558]: return func(*args, **kwargs) [ 710.606419] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.606419] env[62558]: raise e [ 710.606419] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.606419] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 710.606419] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.606419] env[62558]: created_port_ids = self._update_ports_for_instance( [ 710.606419] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.606419] env[62558]: with excutils.save_and_reraise_exception(): [ 710.606419] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.606419] env[62558]: self.force_reraise() [ 710.606419] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.606419] env[62558]: raise self.value [ 710.606419] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.606419] env[62558]: updated_port = self._update_port( [ 710.606419] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.606419] env[62558]: _ensure_no_port_binding_failure(port) [ 710.606419] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.606419] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.607279] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 710.607279] env[62558]: Removing descriptor: 15 [ 710.647847] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.658825] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.659452] env[62558]: ERROR nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Traceback (most recent call last): [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.driver.spawn(context, instance, image_meta, [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] vm_ref = self.build_virtual_machine(instance, [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.659452] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] for vif in network_info: [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return self._sync_wrapper(fn, *args, **kwargs) [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.wait() [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self[:] = self._gt.wait() [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return self._exit_event.wait() [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] current.throw(*self._exc) [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.659823] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] result = function(*args, **kwargs) [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] return func(*args, **kwargs) [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise e [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] nwinfo = self.network_api.allocate_for_instance( [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] created_port_ids = self._update_ports_for_instance( [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] with excutils.save_and_reraise_exception(): [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] self.force_reraise() [ 710.660243] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise self.value [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] updated_port = self._update_port( [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] _ensure_no_port_binding_failure(port) [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] raise exception.PortBindingFailed(port_id=port['id']) [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] nova.exception.PortBindingFailed: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. [ 710.660595] env[62558]: ERROR nova.compute.manager [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] [ 710.660595] env[62558]: DEBUG nova.compute.utils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.661379] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.116s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.663359] env[62558]: INFO nova.compute.claims [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.665859] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Build of instance 65e8971a-a149-4a11-92ad-7c5c8874530c was re-scheduled: Binding failed for port 1672734d-06c0-4100-9b1c-372c9a4c47ec, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.667375] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.667375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.667375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.667375] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.676271] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T05:01:26Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1133726306',id=38,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1424978835',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.676271] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.676271] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.676534] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.676534] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.676534] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.676534] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.676534] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.676705] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.676705] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.676705] env[62558]: DEBUG nova.virt.hardware [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.677559] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd4f337-7edb-40f3-9655-7bfdbe3fe2f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.686967] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2062b518-dddd-4ecc-9e59-5d9d055e1299 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.704164] env[62558]: ERROR nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Traceback (most recent call last): [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] yield resources [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.driver.spawn(context, instance, image_meta, [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] vm_ref = self.build_virtual_machine(instance, [ 710.704164] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] for vif in network_info: [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return self._sync_wrapper(fn, *args, **kwargs) [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.wait() [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self[:] = self._gt.wait() [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return self._exit_event.wait() [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.704597] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] current.throw(*self._exc) [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] result = function(*args, **kwargs) [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return func(*args, **kwargs) [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise e [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] nwinfo = self.network_api.allocate_for_instance( [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] created_port_ids = self._update_ports_for_instance( [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] with excutils.save_and_reraise_exception(): [ 710.704968] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.force_reraise() [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise self.value [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] updated_port = self._update_port( [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] _ensure_no_port_binding_failure(port) [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise exception.PortBindingFailed(port_id=port['id']) [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 710.705340] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] [ 710.705340] env[62558]: INFO nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Terminating instance [ 710.708384] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquiring lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.708384] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquired lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.708384] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.864973] env[62558]: DEBUG oslo_concurrency.lockutils [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] Releasing lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.865901] env[62558]: DEBUG nova.compute.manager [req-1326c15e-ec17-4b55-b690-0407df1d0ac9 req-1731686f-402e-4a32-87ca-7c8ce3a0d47e service nova] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Received event network-vif-deleted-5e28565a-9cb3-4032-97e1-28a94776d683 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.865901] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquired lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.866161] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.194417] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.237854] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.274694] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.365584] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.387257] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.509175] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.780316] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "refresh_cache-65e8971a-a149-4a11-92ad-7c5c8874530c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.780316] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.780316] env[62558]: DEBUG nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.780316] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.797185] env[62558]: DEBUG nova.compute.manager [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Received event network-changed-7df067c2-30b7-4010-8e39-950390e1cf65 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.797381] env[62558]: DEBUG nova.compute.manager [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Refreshing instance network info cache due to event network-changed-7df067c2-30b7-4010-8e39-950390e1cf65. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 711.797585] env[62558]: DEBUG oslo_concurrency.lockutils [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] Acquiring lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.800571] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.868024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Releasing lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.868263] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.868951] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.868951] env[62558]: DEBUG oslo_concurrency.lockutils [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] Acquired lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.868951] env[62558]: DEBUG nova.network.neutron [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Refreshing network info cache for port 7df067c2-30b7-4010-8e39-950390e1cf65 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.870242] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-333d960e-c46f-4f7e-9e91-74c7c579ca86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.883704] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f490a391-0dd7-4ded-8306-0e39f017d4bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.916817] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cecebc6a-25cc-4376-8467-2aa0f909c9bd could not be found. [ 711.920019] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.920019] env[62558]: INFO nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 711.920019] env[62558]: DEBUG oslo.service.loopingcall [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.920379] env[62558]: DEBUG nova.compute.manager [-] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.920379] env[62558]: DEBUG nova.network.neutron [-] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.947738] env[62558]: DEBUG nova.network.neutron [-] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.014531] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Releasing lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.014531] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.014531] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dfe1b17-1aa9-4878-9ac1-3c2b9f048cda {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.023663] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7994fe9c-c3f3-4a86-b94b-fc46388948f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.055696] env[62558]: WARNING nova.virt.vmwareapi.driver [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2ae701c0-3016-4635-8fe1-1e1c314fa748 could not be found. [ 712.055955] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.058884] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43493a82-fa62-46b7-a0c3-624d6f6e0524 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.067859] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a180932-d65a-47ab-ad67-1c6a16f6d5ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.094378] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2ae701c0-3016-4635-8fe1-1e1c314fa748 could not be found. [ 712.094378] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.094378] env[62558]: INFO nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Took 0.08 seconds to destroy the instance on the hypervisor. [ 712.094378] env[62558]: DEBUG oslo.service.loopingcall [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.094378] env[62558]: DEBUG nova.compute.manager [-] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.094378] env[62558]: DEBUG nova.network.neutron [-] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.112798] env[62558]: DEBUG nova.network.neutron [-] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.173942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b19bbe8-66ef-4179-954c-8a48be21cd3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.182853] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873c2942-52bd-4c24-99f3-03a4dd6fe72b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.214381] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8862927a-e828-417d-b8e8-9728e89f9b5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.222676] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7af89a5-4829-418b-915e-f8fb394b9001 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.236584] env[62558]: DEBUG nova.compute.provider_tree [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.246190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.246426] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.304208] env[62558]: DEBUG nova.network.neutron [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.392163] env[62558]: DEBUG nova.network.neutron [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.449217] env[62558]: DEBUG nova.network.neutron [-] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.467590] env[62558]: DEBUG nova.network.neutron [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.615698] env[62558]: DEBUG nova.network.neutron [-] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.740673] env[62558]: DEBUG nova.scheduler.client.report [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.806862] env[62558]: INFO nova.compute.manager [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 65e8971a-a149-4a11-92ad-7c5c8874530c] Took 1.03 seconds to deallocate network for instance. [ 712.952074] env[62558]: INFO nova.compute.manager [-] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Took 1.03 seconds to deallocate network for instance. [ 712.954675] env[62558]: DEBUG nova.compute.claims [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.954877] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.969755] env[62558]: DEBUG oslo_concurrency.lockutils [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] Releasing lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.970118] env[62558]: DEBUG nova.compute.manager [req-7ed1bdb8-300b-45cd-b50c-97d3076dd046 req-26f34e02-18f7-4bfc-9204-8f8da37cabfb service nova] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Received event network-vif-deleted-7df067c2-30b7-4010-8e39-950390e1cf65 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.117863] env[62558]: INFO nova.compute.manager [-] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Took 1.02 seconds to deallocate network for instance. [ 713.246188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.246908] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.253018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.054s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.679150] env[62558]: INFO nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Took 0.56 seconds to detach 1 volumes for instance. [ 713.682080] env[62558]: DEBUG nova.compute.claims [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.682270] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.755683] env[62558]: DEBUG nova.compute.utils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.760128] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.760344] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 713.823620] env[62558]: DEBUG nova.policy [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfa54372b8564929ba1c94f8fdbb4bbe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35e14ae42cc24728a850002f9dab2149', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.832203] env[62558]: INFO nova.scheduler.client.report [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted allocations for instance 65e8971a-a149-4a11-92ad-7c5c8874530c [ 714.157023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be629b89-00a3-4ce3-9b4e-d23a8a126630 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.163625] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2ede09-b9a6-49d8-802f-3f324eb5a630 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.195711] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8031fde-a7c2-446e-a97a-d556e4629672 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.203776] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439f5c47-8773-4525-a84e-36b086909cec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.216823] env[62558]: DEBUG nova.compute.provider_tree [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.260674] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.278658] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Successfully created port: cb291c69-9757-4dd0-98be-704b6b17f8ad {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.342195] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96693efa-31df-4aba-9971-3bde71588dfd tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "65e8971a-a149-4a11-92ad-7c5c8874530c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.002s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.720334] env[62558]: DEBUG nova.scheduler.client.report [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.847065] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.202825] env[62558]: DEBUG nova.compute.manager [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Received event network-changed-cb291c69-9757-4dd0-98be-704b6b17f8ad {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.203040] env[62558]: DEBUG nova.compute.manager [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Refreshing instance network info cache due to event network-changed-cb291c69-9757-4dd0-98be-704b6b17f8ad. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 715.203258] env[62558]: DEBUG oslo_concurrency.lockutils [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] Acquiring lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.203397] env[62558]: DEBUG oslo_concurrency.lockutils [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] Acquired lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.203553] env[62558]: DEBUG nova.network.neutron [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Refreshing network info cache for port cb291c69-9757-4dd0-98be-704b6b17f8ad {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 715.228345] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.228345] env[62558]: ERROR nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] Traceback (most recent call last): [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.driver.spawn(context, instance, image_meta, [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.228345] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] vm_ref = self.build_virtual_machine(instance, [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] for vif in network_info: [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return self._sync_wrapper(fn, *args, **kwargs) [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.wait() [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self[:] = self._gt.wait() [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return self._exit_event.wait() [ 715.228622] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] current.throw(*self._exc) [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] result = function(*args, **kwargs) [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] return func(*args, **kwargs) [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise e [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] nwinfo = self.network_api.allocate_for_instance( [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] created_port_ids = self._update_ports_for_instance( [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.228992] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] with excutils.save_and_reraise_exception(): [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] self.force_reraise() [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise self.value [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] updated_port = self._update_port( [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] _ensure_no_port_binding_failure(port) [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] raise exception.PortBindingFailed(port_id=port['id']) [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] nova.exception.PortBindingFailed: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. [ 715.229343] env[62558]: ERROR nova.compute.manager [instance: b482dcca-ae32-425d-a097-f9af69785835] [ 715.229659] env[62558]: DEBUG nova.compute.utils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 715.229659] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.213s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.233035] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Build of instance b482dcca-ae32-425d-a097-f9af69785835 was re-scheduled: Binding failed for port fc0ee510-649e-46bc-a194-f0c19b5cba74, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 715.233331] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 715.233612] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.233768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.233953] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.270297] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.316762] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.316762] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.316762] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.316997] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.316997] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.316997] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.317860] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.318072] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.318264] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.322048] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.322296] env[62558]: DEBUG nova.virt.hardware [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.323204] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2810a9ce-27d4-4b67-84f9-ad795cb1ba5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.332563] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbaabf9-812d-40b5-9f9d-ca9c280f092d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.378710] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.411254] env[62558]: ERROR nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 715.411254] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.411254] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.411254] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.411254] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.411254] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.411254] env[62558]: ERROR nova.compute.manager raise self.value [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.411254] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 715.411254] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.411254] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 715.411961] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.411961] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 715.411961] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 715.411961] env[62558]: ERROR nova.compute.manager [ 715.411961] env[62558]: Traceback (most recent call last): [ 715.411961] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 715.411961] env[62558]: listener.cb(fileno) [ 715.411961] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.411961] env[62558]: result = function(*args, **kwargs) [ 715.411961] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.411961] env[62558]: return func(*args, **kwargs) [ 715.411961] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.411961] env[62558]: raise e [ 715.411961] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.411961] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 715.411961] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.411961] env[62558]: created_port_ids = self._update_ports_for_instance( [ 715.411961] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.411961] env[62558]: with excutils.save_and_reraise_exception(): [ 715.411961] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.411961] env[62558]: self.force_reraise() [ 715.411961] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.411961] env[62558]: raise self.value [ 715.411961] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.411961] env[62558]: updated_port = self._update_port( [ 715.411961] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.411961] env[62558]: _ensure_no_port_binding_failure(port) [ 715.411961] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.411961] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 715.413247] env[62558]: nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 715.413247] env[62558]: Removing descriptor: 22 [ 715.413247] env[62558]: ERROR nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Traceback (most recent call last): [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] yield resources [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.driver.spawn(context, instance, image_meta, [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.413247] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] vm_ref = self.build_virtual_machine(instance, [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] for vif in network_info: [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self._sync_wrapper(fn, *args, **kwargs) [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.wait() [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self[:] = self._gt.wait() [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self._exit_event.wait() [ 715.414082] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] result = hub.switch() [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self.greenlet.switch() [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] result = function(*args, **kwargs) [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return func(*args, **kwargs) [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise e [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] nwinfo = self.network_api.allocate_for_instance( [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 715.414730] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] created_port_ids = self._update_ports_for_instance( [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] with excutils.save_and_reraise_exception(): [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.force_reraise() [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise self.value [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] updated_port = self._update_port( [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] _ensure_no_port_binding_failure(port) [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.415199] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise exception.PortBindingFailed(port_id=port['id']) [ 715.415527] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 715.415527] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] [ 715.415527] env[62558]: INFO nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Terminating instance [ 715.416946] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquiring lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.720181] env[62558]: DEBUG nova.network.neutron [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.754418] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.837479] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.862139] env[62558]: DEBUG nova.network.neutron [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.098579] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e09776-f198-4699-8e33-7f9280196c87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.107675] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf29651-88ad-4f0e-bc7a-4a369b2f7dfa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.138376] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7989d8be-197e-4a64-aa50-6cfe3100d07f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.146206] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e94b21d-f729-441c-9aec-5fe545304c8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.160020] env[62558]: DEBUG nova.compute.provider_tree [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.345246] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-b482dcca-ae32-425d-a097-f9af69785835" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.345547] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.345680] env[62558]: DEBUG nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.345842] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.366642] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.368350] env[62558]: DEBUG oslo_concurrency.lockutils [req-bcec1a91-01a0-4fdb-9487-afb0bde9889c req-aa0a5d45-0ea3-42bb-83f5-9584e6a2122e service nova] Releasing lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.368944] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquired lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.369132] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.663176] env[62558]: DEBUG nova.scheduler.client.report [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.748568] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.748823] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.869503] env[62558]: DEBUG nova.network.neutron [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.889497] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.002307] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.168595] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.169312] env[62558]: ERROR nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] Traceback (most recent call last): [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.driver.spawn(context, instance, image_meta, [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] vm_ref = self.build_virtual_machine(instance, [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.169312] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] for vif in network_info: [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self._sync_wrapper(fn, *args, **kwargs) [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.wait() [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self[:] = self._gt.wait() [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self._exit_event.wait() [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] result = hub.switch() [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.169701] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return self.greenlet.switch() [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] result = function(*args, **kwargs) [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] return func(*args, **kwargs) [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise e [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] nwinfo = self.network_api.allocate_for_instance( [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] created_port_ids = self._update_ports_for_instance( [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] with excutils.save_and_reraise_exception(): [ 717.170150] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] self.force_reraise() [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise self.value [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] updated_port = self._update_port( [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] _ensure_no_port_binding_failure(port) [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] raise exception.PortBindingFailed(port_id=port['id']) [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] nova.exception.PortBindingFailed: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. [ 717.170551] env[62558]: ERROR nova.compute.manager [instance: f8db751a-c495-467c-ae32-addb4387346d] [ 717.170893] env[62558]: DEBUG nova.compute.utils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.171277] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.638s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.174275] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Build of instance f8db751a-c495-467c-ae32-addb4387346d was re-scheduled: Binding failed for port 725c06cc-3cf6-4685-bfd8-40182e6b717e, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.174710] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.174935] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquiring lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.175101] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Acquired lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.175269] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.248766] env[62558]: DEBUG nova.compute.manager [req-30a48a2b-1784-4818-bc8d-7719ad09ac74 req-2f7a305b-e424-4776-b94d-f656c4512b5c service nova] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Received event network-vif-deleted-cb291c69-9757-4dd0-98be-704b6b17f8ad {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.373524] env[62558]: INFO nova.compute.manager [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: b482dcca-ae32-425d-a097-f9af69785835] Took 1.03 seconds to deallocate network for instance. [ 717.505264] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Releasing lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.505691] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.505904] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 717.506223] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10a6af8d-daf7-4996-8a6b-ccb702e191d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.515779] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3455e502-2e34-4b8a-afe0-6351f29daa33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.538285] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a092fb50-c750-4b91-a94e-e11e9eb7faf5 could not be found. [ 717.538387] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.538578] env[62558]: INFO nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 717.538872] env[62558]: DEBUG oslo.service.loopingcall [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.539397] env[62558]: DEBUG nova.compute.manager [-] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.539494] env[62558]: DEBUG nova.network.neutron [-] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.570845] env[62558]: DEBUG nova.network.neutron [-] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.694978] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.789730] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.007763] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9727ebac-7600-4276-84bc-6f864b7dd143 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.015146] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-465ed293-dcc9-4d32-a844-385abe19c094 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.046108] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fccdf77-a35c-4f36-9834-6f1fc35fabc0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.053792] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7280e2ef-1aaf-4f8e-83c1-4dee900b34e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.066576] env[62558]: DEBUG nova.compute.provider_tree [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.073348] env[62558]: DEBUG nova.network.neutron [-] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.293451] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Releasing lock "refresh_cache-f8db751a-c495-467c-ae32-addb4387346d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.293750] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.293946] env[62558]: DEBUG nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.294131] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.307056] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.400230] env[62558]: INFO nova.scheduler.client.report [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted allocations for instance b482dcca-ae32-425d-a097-f9af69785835 [ 718.569779] env[62558]: DEBUG nova.scheduler.client.report [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.577296] env[62558]: INFO nova.compute.manager [-] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Took 1.04 seconds to deallocate network for instance. [ 718.579788] env[62558]: DEBUG nova.compute.claims [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 718.579788] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.809519] env[62558]: DEBUG nova.network.neutron [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.908266] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f8bf109-00d0-4ecb-8870-f41f357cec1c tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "b482dcca-ae32-425d-a097-f9af69785835" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.751s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.076163] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.076912] env[62558]: ERROR nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Traceback (most recent call last): [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.driver.spawn(context, instance, image_meta, [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] vm_ref = self.build_virtual_machine(instance, [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.076912] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] for vif in network_info: [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return self._sync_wrapper(fn, *args, **kwargs) [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.wait() [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self[:] = self._gt.wait() [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return self._exit_event.wait() [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] current.throw(*self._exc) [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.077291] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] result = function(*args, **kwargs) [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] return func(*args, **kwargs) [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise e [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] nwinfo = self.network_api.allocate_for_instance( [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] created_port_ids = self._update_ports_for_instance( [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] with excutils.save_and_reraise_exception(): [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] self.force_reraise() [ 719.077700] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise self.value [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] updated_port = self._update_port( [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] _ensure_no_port_binding_failure(port) [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] raise exception.PortBindingFailed(port_id=port['id']) [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] nova.exception.PortBindingFailed: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. [ 719.078222] env[62558]: ERROR nova.compute.manager [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] [ 719.078222] env[62558]: DEBUG nova.compute.utils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.078999] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.086s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.081869] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Build of instance 9b49513e-4c01-4664-be2c-6058a357b0db was re-scheduled: Binding failed for port 52f07dbe-eb8d-4bd8-948e-874e5d745201, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.082310] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.082530] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquiring lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.082672] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Acquired lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.082826] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.312443] env[62558]: INFO nova.compute.manager [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] [instance: f8db751a-c495-467c-ae32-addb4387346d] Took 1.02 seconds to deallocate network for instance. [ 719.411017] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.602110] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.690225] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.697288] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.697503] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.908469] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde6dc84-c841-470f-a82f-ccfc3e4c4437 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.918433] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecee27e-0233-4106-a17c-6355587be39c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.951574] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.952372] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a892c687-f914-4cb7-b832-f56a220755a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.960403] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1b2663-59af-464f-ae6d-1fd3d37b238a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.974199] env[62558]: DEBUG nova.compute.provider_tree [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.194959] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Releasing lock "refresh_cache-9b49513e-4c01-4664-be2c-6058a357b0db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.195233] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.195415] env[62558]: DEBUG nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.195581] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.218450] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.341465] env[62558]: INFO nova.scheduler.client.report [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Deleted allocations for instance f8db751a-c495-467c-ae32-addb4387346d [ 720.478837] env[62558]: DEBUG nova.scheduler.client.report [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.720895] env[62558]: DEBUG nova.network.neutron [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.850362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c54f3635-758b-448b-bdab-43f33e8c4599 tempest-ImagesOneServerNegativeTestJSON-2144231741 tempest-ImagesOneServerNegativeTestJSON-2144231741-project-member] Lock "f8db751a-c495-467c-ae32-addb4387346d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.245s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.985774] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.986472] env[62558]: ERROR nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Traceback (most recent call last): [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.driver.spawn(context, instance, image_meta, [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] vm_ref = self.build_virtual_machine(instance, [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.986472] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] for vif in network_info: [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return self._sync_wrapper(fn, *args, **kwargs) [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.wait() [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self[:] = self._gt.wait() [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return self._exit_event.wait() [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] current.throw(*self._exc) [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.986826] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] result = function(*args, **kwargs) [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] return func(*args, **kwargs) [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise e [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] nwinfo = self.network_api.allocate_for_instance( [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] created_port_ids = self._update_ports_for_instance( [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] with excutils.save_and_reraise_exception(): [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] self.force_reraise() [ 720.987238] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise self.value [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] updated_port = self._update_port( [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] _ensure_no_port_binding_failure(port) [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] raise exception.PortBindingFailed(port_id=port['id']) [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] nova.exception.PortBindingFailed: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. [ 720.987638] env[62558]: ERROR nova.compute.manager [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] [ 720.987638] env[62558]: DEBUG nova.compute.utils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.988375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.570s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.990295] env[62558]: INFO nova.compute.claims [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.992719] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Build of instance 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4 was re-scheduled: Binding failed for port a56124a9-37f0-4760-9955-84b74f5385df, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 720.993133] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 720.993355] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.993497] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.993651] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.223231] env[62558]: INFO nova.compute.manager [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] [instance: 9b49513e-4c01-4664-be2c-6058a357b0db] Took 1.03 seconds to deallocate network for instance. [ 721.353199] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.521118] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.655638] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.875067] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.158015] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.158275] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.158453] env[62558]: DEBUG nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.158616] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.179362] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.253812] env[62558]: INFO nova.scheduler.client.report [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Deleted allocations for instance 9b49513e-4c01-4664-be2c-6058a357b0db [ 722.360480] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2128b2b0-5abe-4391-bbe2-a005bcfbfa9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.368042] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2005b39-5afa-4828-b836-fbaf50d57702 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.397731] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f450dbed-4e7d-4346-aa2b-0e01506cbbc8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.404970] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f5a0be-5a55-4056-aed8-636bc1e5abde {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.419196] env[62558]: DEBUG nova.compute.provider_tree [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.682396] env[62558]: DEBUG nova.network.neutron [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.766871] env[62558]: DEBUG oslo_concurrency.lockutils [None req-40d13496-3d76-4939-84d1-65585c7bf3ae tempest-ServersTestJSON-1213459211 tempest-ServersTestJSON-1213459211-project-member] Lock "9b49513e-4c01-4664-be2c-6058a357b0db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.671s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.921837] env[62558]: DEBUG nova.scheduler.client.report [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.187598] env[62558]: INFO nova.compute.manager [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4] Took 1.03 seconds to deallocate network for instance. [ 723.272205] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.430018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.430018] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.431390] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.421s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.793779] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.937981] env[62558]: DEBUG nova.compute.utils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.943073] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.943366] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.144715] env[62558]: DEBUG nova.policy [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a14f2591414344febb2203f864bd4402', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74119b04be494ed7ba03cd3205b1e0d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.229563] env[62558]: INFO nova.scheduler.client.report [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance 7c22d751-2aa8-4794-9fbb-3be5c1ac40f4 [ 724.357039] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e4dc7b-6c36-4896-97a6-79a676c0de72 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.368686] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1685e628-4da2-43e1-9f6d-433d0517b96b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.405074] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388739bc-3583-4315-8123-7eeafe0b29c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.413337] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbe73ac-6528-4b94-be92-7982f0ce1475 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.427196] env[62558]: DEBUG nova.compute.provider_tree [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.441582] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.647218] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Successfully created port: 06567788-b01a-48a6-a6c0-27fa9897a25f {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.743038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-cc7fa442-36b0-4975-84a6-f00ccf046ddb tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "7c22d751-2aa8-4794-9fbb-3be5c1ac40f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.954s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.932789] env[62558]: DEBUG nova.scheduler.client.report [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.249258] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.440262] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.440834] env[62558]: ERROR nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Traceback (most recent call last): [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.driver.spawn(context, instance, image_meta, [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] vm_ref = self.build_virtual_machine(instance, [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.440834] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] for vif in network_info: [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return self._sync_wrapper(fn, *args, **kwargs) [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.wait() [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self[:] = self._gt.wait() [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return self._exit_event.wait() [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] current.throw(*self._exc) [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.441485] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] result = function(*args, **kwargs) [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] return func(*args, **kwargs) [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise e [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] nwinfo = self.network_api.allocate_for_instance( [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] created_port_ids = self._update_ports_for_instance( [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] with excutils.save_and_reraise_exception(): [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] self.force_reraise() [ 725.441967] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise self.value [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] updated_port = self._update_port( [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] _ensure_no_port_binding_failure(port) [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] raise exception.PortBindingFailed(port_id=port['id']) [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] nova.exception.PortBindingFailed: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. [ 725.442994] env[62558]: ERROR nova.compute.manager [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] [ 725.442994] env[62558]: DEBUG nova.compute.utils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.443758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.488s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.450861] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Build of instance 2888fde6-8a1b-41ce-a482-cb99acd74a6e was re-scheduled: Binding failed for port 892b4af7-fadf-4cfb-817a-3e6afbf89fab, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.451447] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.451664] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.451811] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.452015] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.453813] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.495397] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.495894] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.495894] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.495997] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.496147] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.497595] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.498046] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.498288] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.498738] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.498993] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.499219] env[62558]: DEBUG nova.virt.hardware [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.500620] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63e6b4d-97f4-47bc-a01c-ae016ba461ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.511320] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f0d72e-b8c5-4253-8201-6a120a3f5528 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.673016] env[62558]: DEBUG nova.compute.manager [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Received event network-changed-06567788-b01a-48a6-a6c0-27fa9897a25f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.673173] env[62558]: DEBUG nova.compute.manager [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Refreshing instance network info cache due to event network-changed-06567788-b01a-48a6-a6c0-27fa9897a25f. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.674013] env[62558]: DEBUG oslo_concurrency.lockutils [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] Acquiring lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.674013] env[62558]: DEBUG oslo_concurrency.lockutils [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] Acquired lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.674013] env[62558]: DEBUG nova.network.neutron [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Refreshing network info cache for port 06567788-b01a-48a6-a6c0-27fa9897a25f {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.773552] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.991531] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.026767] env[62558]: ERROR nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 726.026767] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.026767] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.026767] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.026767] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.026767] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.026767] env[62558]: ERROR nova.compute.manager raise self.value [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.026767] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.026767] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.026767] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.027275] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.027275] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.027275] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 726.027275] env[62558]: ERROR nova.compute.manager [ 726.027275] env[62558]: Traceback (most recent call last): [ 726.027275] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.027275] env[62558]: listener.cb(fileno) [ 726.027275] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.027275] env[62558]: result = function(*args, **kwargs) [ 726.027275] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.027275] env[62558]: return func(*args, **kwargs) [ 726.027275] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.027275] env[62558]: raise e [ 726.027275] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.027275] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 726.027275] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.027275] env[62558]: created_port_ids = self._update_ports_for_instance( [ 726.027275] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.027275] env[62558]: with excutils.save_and_reraise_exception(): [ 726.027275] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.027275] env[62558]: self.force_reraise() [ 726.027275] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.027275] env[62558]: raise self.value [ 726.027275] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.027275] env[62558]: updated_port = self._update_port( [ 726.027275] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.027275] env[62558]: _ensure_no_port_binding_failure(port) [ 726.027275] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.027275] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.028054] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 726.028054] env[62558]: Removing descriptor: 22 [ 726.028054] env[62558]: ERROR nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Traceback (most recent call last): [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] yield resources [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.driver.spawn(context, instance, image_meta, [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.028054] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] vm_ref = self.build_virtual_machine(instance, [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] for vif in network_info: [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self._sync_wrapper(fn, *args, **kwargs) [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.wait() [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self[:] = self._gt.wait() [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self._exit_event.wait() [ 726.028408] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] result = hub.switch() [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self.greenlet.switch() [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] result = function(*args, **kwargs) [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return func(*args, **kwargs) [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise e [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] nwinfo = self.network_api.allocate_for_instance( [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.028805] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] created_port_ids = self._update_ports_for_instance( [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] with excutils.save_and_reraise_exception(): [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.force_reraise() [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise self.value [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] updated_port = self._update_port( [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] _ensure_no_port_binding_failure(port) [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.029264] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise exception.PortBindingFailed(port_id=port['id']) [ 726.029628] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 726.029628] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] [ 726.029628] env[62558]: INFO nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Terminating instance [ 726.030267] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.105247] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.194307] env[62558]: DEBUG nova.network.neutron [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.270932] env[62558]: DEBUG nova.network.neutron [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.375482] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea123e6-577e-41c0-af35-2149de848388 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.383102] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3eed79-7d7c-4f94-95df-79742f28d643 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.420332] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb0bf90-b309-4d9b-b0ac-578b732c40c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.429259] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac8ae84-4f4d-4797-868b-1dc61c019141 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.444277] env[62558]: DEBUG nova.compute.provider_tree [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.610824] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-2888fde6-8a1b-41ce-a482-cb99acd74a6e" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.611091] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.611280] env[62558]: DEBUG nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.611442] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.633476] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.780018] env[62558]: DEBUG oslo_concurrency.lockutils [req-23800559-30fa-4a3d-817f-b21a43e8d9dd req-26d29230-0898-4329-a184-aabc3528f8c0 service nova] Releasing lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.780018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquired lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.780018] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.947945] env[62558]: DEBUG nova.scheduler.client.report [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.139902] env[62558]: DEBUG nova.network.neutron [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.304483] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.357079] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.457928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.458763] env[62558]: ERROR nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Traceback (most recent call last): [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.driver.spawn(context, instance, image_meta, [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] vm_ref = self.build_virtual_machine(instance, [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.458763] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] for vif in network_info: [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return self._sync_wrapper(fn, *args, **kwargs) [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.wait() [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self[:] = self._gt.wait() [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return self._exit_event.wait() [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] current.throw(*self._exc) [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.459214] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] result = function(*args, **kwargs) [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] return func(*args, **kwargs) [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise e [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] nwinfo = self.network_api.allocate_for_instance( [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] created_port_ids = self._update_ports_for_instance( [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] with excutils.save_and_reraise_exception(): [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] self.force_reraise() [ 727.459693] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise self.value [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] updated_port = self._update_port( [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] _ensure_no_port_binding_failure(port) [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] raise exception.PortBindingFailed(port_id=port['id']) [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] nova.exception.PortBindingFailed: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. [ 727.460133] env[62558]: ERROR nova.compute.manager [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] [ 727.460133] env[62558]: DEBUG nova.compute.utils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.460759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.778s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.463776] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Build of instance cecebc6a-25cc-4376-8467-2aa0f909c9bd was re-scheduled: Binding failed for port 7df067c2-30b7-4010-8e39-950390e1cf65, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 727.464305] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 727.464531] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquiring lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.464674] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Acquired lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.464828] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.644294] env[62558]: INFO nova.compute.manager [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 2888fde6-8a1b-41ce-a482-cb99acd74a6e] Took 1.03 seconds to deallocate network for instance. [ 727.715747] env[62558]: DEBUG nova.compute.manager [req-af0ba434-ce5f-4bc3-8977-4927f8bde5d1 req-0fcd8ef6-cee0-40d1-8c95-337227313b8d service nova] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Received event network-vif-deleted-06567788-b01a-48a6-a6c0-27fa9897a25f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.856063] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.857120] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.861075] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Releasing lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.861589] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.862031] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.862510] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2eaa61b2-0608-400c-9938-afdca1509b55 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.874717] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55057c8f-9142-4860-8c31-856d1f9fb240 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.902232] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 151768fd-c2c8-491f-93b5-bbc53bd69e2c could not be found. [ 727.902681] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.902681] env[62558]: INFO nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 727.902944] env[62558]: DEBUG oslo.service.loopingcall [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.903118] env[62558]: DEBUG nova.compute.manager [-] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.903212] env[62558]: DEBUG nova.network.neutron [-] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.920965] env[62558]: DEBUG nova.network.neutron [-] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.987957] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.069285] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.426656] env[62558]: DEBUG nova.network.neutron [-] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.437152] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d384e3-12c4-4f03-b29f-8f3193a4d82a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.445955] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbdefd5-eeec-432c-a3a5-633a957ea7b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.479797] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16b7ef9-36b7-4df7-b992-63e964090cd6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.487174] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01c7cd7-6392-49a3-b3f8-23de49c350d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.508330] env[62558]: DEBUG nova.compute.provider_tree [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.571363] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Releasing lock "refresh_cache-cecebc6a-25cc-4376-8467-2aa0f909c9bd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.574500] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 728.574500] env[62558]: DEBUG nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.574500] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.595108] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.684732] env[62558]: INFO nova.scheduler.client.report [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance 2888fde6-8a1b-41ce-a482-cb99acd74a6e [ 728.930900] env[62558]: INFO nova.compute.manager [-] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Took 1.03 seconds to deallocate network for instance. [ 728.933776] env[62558]: DEBUG nova.compute.claims [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.933976] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.011281] env[62558]: DEBUG nova.scheduler.client.report [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.091386] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.091637] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.100900] env[62558]: DEBUG nova.network.neutron [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.192839] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d42d51c7-65db-48e7-bc7a-04254531c924 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "2888fde6-8a1b-41ce-a482-cb99acd74a6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.195s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.516462] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.517116] env[62558]: ERROR nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Traceback (most recent call last): [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.driver.spawn(context, instance, image_meta, [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] vm_ref = self.build_virtual_machine(instance, [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.517116] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] for vif in network_info: [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return self._sync_wrapper(fn, *args, **kwargs) [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.wait() [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self[:] = self._gt.wait() [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return self._exit_event.wait() [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] current.throw(*self._exc) [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.517466] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] result = function(*args, **kwargs) [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] return func(*args, **kwargs) [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise e [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] nwinfo = self.network_api.allocate_for_instance( [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] created_port_ids = self._update_ports_for_instance( [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] with excutils.save_and_reraise_exception(): [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] self.force_reraise() [ 729.517827] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise self.value [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] updated_port = self._update_port( [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] _ensure_no_port_binding_failure(port) [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] raise exception.PortBindingFailed(port_id=port['id']) [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] nova.exception.PortBindingFailed: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. [ 729.518194] env[62558]: ERROR nova.compute.manager [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] [ 729.518194] env[62558]: DEBUG nova.compute.utils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.519095] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.140s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.520511] env[62558]: INFO nova.compute.claims [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.523086] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Build of instance 2ae701c0-3016-4635-8fe1-1e1c314fa748 was re-scheduled: Binding failed for port 5e28565a-9cb3-4032-97e1-28a94776d683, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 729.523701] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 729.523744] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquiring lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.523870] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Acquired lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.524034] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.605315] env[62558]: INFO nova.compute.manager [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] [instance: cecebc6a-25cc-4376-8467-2aa0f909c9bd] Took 1.03 seconds to deallocate network for instance. [ 729.695728] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.046809] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.138301] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.147302] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.147449] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.228588] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.637097] env[62558]: INFO nova.scheduler.client.report [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Deleted allocations for instance cecebc6a-25cc-4376-8467-2aa0f909c9bd [ 730.649445] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Releasing lock "refresh_cache-2ae701c0-3016-4635-8fe1-1e1c314fa748" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.649669] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 730.649846] env[62558]: DEBUG nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.651114] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.670398] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.944849] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dee27c5-054a-470a-a5af-97c441a17154 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.954261] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e6f771-be7c-41ab-8f41-a1816724e354 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.988680] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4736e67-abd2-410e-b1f4-0fcb1f3826de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.997710] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98202a37-025b-4915-9e97-0a63f1fce920 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.012131] env[62558]: DEBUG nova.compute.provider_tree [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.155068] env[62558]: DEBUG oslo_concurrency.lockutils [None req-daee43df-5813-49a6-8cf4-15c70a32bfe1 tempest-ServersWithSpecificFlavorTestJSON-1103641373 tempest-ServersWithSpecificFlavorTestJSON-1103641373-project-member] Lock "cecebc6a-25cc-4376-8467-2aa0f909c9bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.781s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.174172] env[62558]: DEBUG nova.network.neutron [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.515709] env[62558]: DEBUG nova.scheduler.client.report [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.609673] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.609906] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.657640] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 731.676824] env[62558]: INFO nova.compute.manager [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] [instance: 2ae701c0-3016-4635-8fe1-1e1c314fa748] Took 1.03 seconds to deallocate network for instance. [ 732.022023] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.023120] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.025247] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.445s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.181388] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.530252] env[62558]: DEBUG nova.compute.utils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.534898] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.535084] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.593818] env[62558]: DEBUG nova.policy [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a14f2591414344febb2203f864bd4402', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74119b04be494ed7ba03cd3205b1e0d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.716719] env[62558]: INFO nova.scheduler.client.report [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Deleted allocations for instance 2ae701c0-3016-4635-8fe1-1e1c314fa748 [ 732.911691] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Successfully created port: c027a1a0-5707-4c6b-9967-6c4973090fde {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.940625] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.940625] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.962500] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d213ff-4335-4da8-ac93-d42286d95c0b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.974088] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fce735-e1f2-441d-b665-170290aef8c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.007843] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a686b05-09f4-4483-bb0f-06426037061a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.015394] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf037c51-db3e-4b0d-b3ad-c6cd1ec228a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.029204] env[62558]: DEBUG nova.compute.provider_tree [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.036083] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.230191] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e734ff10-c636-42f3-958f-5c148d703c68 tempest-ServerActionsV293TestJSON-1843138847 tempest-ServerActionsV293TestJSON-1843138847-project-member] Lock "2ae701c0-3016-4635-8fe1-1e1c314fa748" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.029s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.446523] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.446695] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 733.446813] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 733.534026] env[62558]: DEBUG nova.scheduler.client.report [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.646709] env[62558]: DEBUG nova.compute.manager [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Received event network-changed-c027a1a0-5707-4c6b-9967-6c4973090fde {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.646861] env[62558]: DEBUG nova.compute.manager [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Refreshing instance network info cache due to event network-changed-c027a1a0-5707-4c6b-9967-6c4973090fde. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.647861] env[62558]: DEBUG oslo_concurrency.lockutils [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] Acquiring lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.647861] env[62558]: DEBUG oslo_concurrency.lockutils [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] Acquired lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.647861] env[62558]: DEBUG nova.network.neutron [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Refreshing network info cache for port c027a1a0-5707-4c6b-9967-6c4973090fde {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.733542] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.819821] env[62558]: ERROR nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 733.819821] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.819821] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.819821] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.819821] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.819821] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.819821] env[62558]: ERROR nova.compute.manager raise self.value [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.819821] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.819821] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.819821] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.820438] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.820438] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.820438] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 733.820438] env[62558]: ERROR nova.compute.manager [ 733.820438] env[62558]: Traceback (most recent call last): [ 733.820438] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.820438] env[62558]: listener.cb(fileno) [ 733.820438] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.820438] env[62558]: result = function(*args, **kwargs) [ 733.820438] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.820438] env[62558]: return func(*args, **kwargs) [ 733.820438] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.820438] env[62558]: raise e [ 733.820438] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.820438] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 733.820438] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.820438] env[62558]: created_port_ids = self._update_ports_for_instance( [ 733.820438] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.820438] env[62558]: with excutils.save_and_reraise_exception(): [ 733.820438] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.820438] env[62558]: self.force_reraise() [ 733.820438] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.820438] env[62558]: raise self.value [ 733.820438] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.820438] env[62558]: updated_port = self._update_port( [ 733.820438] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.820438] env[62558]: _ensure_no_port_binding_failure(port) [ 733.820438] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.820438] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.821369] env[62558]: nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 733.821369] env[62558]: Removing descriptor: 15 [ 733.950641] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 733.951175] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 733.951556] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Didn't find any instances for network info cache update. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 733.951817] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.952782] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.952863] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.953135] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.953434] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.953728] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 733.953893] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 733.954084] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.037149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.037779] env[62558]: ERROR nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Traceback (most recent call last): [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.driver.spawn(context, instance, image_meta, [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] vm_ref = self.build_virtual_machine(instance, [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.037779] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] for vif in network_info: [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self._sync_wrapper(fn, *args, **kwargs) [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.wait() [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self[:] = self._gt.wait() [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self._exit_event.wait() [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] result = hub.switch() [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.038172] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return self.greenlet.switch() [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] result = function(*args, **kwargs) [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] return func(*args, **kwargs) [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise e [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] nwinfo = self.network_api.allocate_for_instance( [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] created_port_ids = self._update_ports_for_instance( [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] with excutils.save_and_reraise_exception(): [ 734.038556] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] self.force_reraise() [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise self.value [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] updated_port = self._update_port( [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] _ensure_no_port_binding_failure(port) [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] raise exception.PortBindingFailed(port_id=port['id']) [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] nova.exception.PortBindingFailed: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. [ 734.038934] env[62558]: ERROR nova.compute.manager [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] [ 734.039299] env[62558]: DEBUG nova.compute.utils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.039787] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.088s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.041705] env[62558]: INFO nova.compute.claims [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.044307] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Build of instance a092fb50-c750-4b91-a94e-e11e9eb7faf5 was re-scheduled: Binding failed for port cb291c69-9757-4dd0-98be-704b6b17f8ad, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.044751] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.044972] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquiring lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.045129] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Acquired lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.045286] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.046790] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.076205] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.076465] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.076623] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.076803] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.076944] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.079546] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.082894] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.082894] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.082894] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.082894] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.082894] env[62558]: DEBUG nova.virt.hardware [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.083297] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b38d912-f6b4-44db-8af0-3319c2666955 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.095717] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ea8b4f-e9c9-4f9a-becd-f119f4c6eca1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.113377] env[62558]: ERROR nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Traceback (most recent call last): [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] yield resources [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.driver.spawn(context, instance, image_meta, [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] vm_ref = self.build_virtual_machine(instance, [ 734.113377] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] for vif in network_info: [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return self._sync_wrapper(fn, *args, **kwargs) [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.wait() [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self[:] = self._gt.wait() [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return self._exit_event.wait() [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.113917] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] current.throw(*self._exc) [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] result = function(*args, **kwargs) [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return func(*args, **kwargs) [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise e [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] nwinfo = self.network_api.allocate_for_instance( [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] created_port_ids = self._update_ports_for_instance( [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] with excutils.save_and_reraise_exception(): [ 734.114454] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.force_reraise() [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise self.value [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] updated_port = self._update_port( [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] _ensure_no_port_binding_failure(port) [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise exception.PortBindingFailed(port_id=port['id']) [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 734.114916] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] [ 734.114916] env[62558]: INFO nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Terminating instance [ 734.116146] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.177269] env[62558]: DEBUG nova.network.neutron [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.259329] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.267950] env[62558]: DEBUG nova.network.neutron [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.458171] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.572740] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.694594] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.771614] env[62558]: DEBUG oslo_concurrency.lockutils [req-39f569a9-8f67-44d0-80c8-6978982ae3e6 req-eab33f5e-6124-4e85-ad57-d0515ec7b8e1 service nova] Releasing lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.772296] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquired lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.772296] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.196547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Releasing lock "refresh_cache-a092fb50-c750-4b91-a94e-e11e9eb7faf5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.196858] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.196961] env[62558]: DEBUG nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.197145] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.224783] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.315423] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.408410] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.450804] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3018a36-8c42-4147-9601-db604ddf43e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.458965] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1d0856-8fb5-4152-be02-3182e7e1e905 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.493654] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5950e3ae-c358-4771-974d-5845a151bac0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.501575] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e111487-8860-4908-9df1-9a0a28b56941 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.515306] env[62558]: DEBUG nova.compute.provider_tree [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.728492] env[62558]: DEBUG nova.network.neutron [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.731886] env[62558]: DEBUG nova.compute.manager [req-724226fc-5d37-419a-b8aa-6609a8fdef4d req-0e1cff10-b6d8-4d83-a90f-8d3b70c25b5f service nova] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Received event network-vif-deleted-c027a1a0-5707-4c6b-9967-6c4973090fde {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.914017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Releasing lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.914017] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.914017] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.914017] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77d4420c-5f82-42f3-a761-9c468b74bfd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.922731] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c224c3-6178-406c-a949-ab58f2e93764 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.946449] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f3128303-9b2c-4f14-9ea0-093f55ecb063 could not be found. [ 735.946706] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.946888] env[62558]: INFO nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Took 0.03 seconds to destroy the instance on the hypervisor. [ 735.947167] env[62558]: DEBUG oslo.service.loopingcall [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.947378] env[62558]: DEBUG nova.compute.manager [-] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.947469] env[62558]: DEBUG nova.network.neutron [-] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.018667] env[62558]: DEBUG nova.scheduler.client.report [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.080863] env[62558]: DEBUG nova.network.neutron [-] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.232999] env[62558]: INFO nova.compute.manager [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] [instance: a092fb50-c750-4b91-a94e-e11e9eb7faf5] Took 1.04 seconds to deallocate network for instance. [ 736.525352] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.525940] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.529400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.654s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.531017] env[62558]: INFO nova.compute.claims [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.583364] env[62558]: DEBUG nova.network.neutron [-] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.037165] env[62558]: DEBUG nova.compute.utils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.040850] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 737.041038] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.085754] env[62558]: INFO nova.compute.manager [-] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Took 1.14 seconds to deallocate network for instance. [ 737.087930] env[62558]: DEBUG nova.compute.claims [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.088154] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.090042] env[62558]: DEBUG nova.policy [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c28a344dd1340229e157824595f2eaf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5cc4e04eb5ed47799e45a2ca0b1676dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.265825] env[62558]: INFO nova.scheduler.client.report [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Deleted allocations for instance a092fb50-c750-4b91-a94e-e11e9eb7faf5 [ 737.385017] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Successfully created port: 488fd776-f368-4d06-abf6-03c8a8880f86 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.543975] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.775543] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e779a3d0-d63a-40eb-85d1-89c7aaeb1b6a tempest-ServersTestFqdnHostnames-1298339553 tempest-ServersTestFqdnHostnames-1298339553-project-member] Lock "a092fb50-c750-4b91-a94e-e11e9eb7faf5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.217s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.940023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a2929e-b1d1-41f2-a319-00a0d9e0d4c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.949052] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547c2f7b-a148-4346-a31e-a2fc1b0d19b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.983947] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63ffa5b-7553-49bf-97fc-5c0066c7b14e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.993576] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3887587-7f6c-4672-9fd3-da67a3606879 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.008244] env[62558]: DEBUG nova.compute.provider_tree [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.148547] env[62558]: DEBUG nova.compute.manager [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Received event network-changed-488fd776-f368-4d06-abf6-03c8a8880f86 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.148878] env[62558]: DEBUG nova.compute.manager [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Refreshing instance network info cache due to event network-changed-488fd776-f368-4d06-abf6-03c8a8880f86. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.149141] env[62558]: DEBUG oslo_concurrency.lockutils [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] Acquiring lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.149328] env[62558]: DEBUG oslo_concurrency.lockutils [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] Acquired lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.149534] env[62558]: DEBUG nova.network.neutron [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Refreshing network info cache for port 488fd776-f368-4d06-abf6-03c8a8880f86 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.284900] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.317196] env[62558]: ERROR nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 738.317196] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.317196] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.317196] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.317196] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.317196] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.317196] env[62558]: ERROR nova.compute.manager raise self.value [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.317196] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.317196] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.317196] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.317730] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.317730] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.317730] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 738.317730] env[62558]: ERROR nova.compute.manager [ 738.317730] env[62558]: Traceback (most recent call last): [ 738.317730] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.317730] env[62558]: listener.cb(fileno) [ 738.317730] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.317730] env[62558]: result = function(*args, **kwargs) [ 738.317730] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.317730] env[62558]: return func(*args, **kwargs) [ 738.317730] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.317730] env[62558]: raise e [ 738.317730] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.317730] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 738.317730] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.317730] env[62558]: created_port_ids = self._update_ports_for_instance( [ 738.317730] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.317730] env[62558]: with excutils.save_and_reraise_exception(): [ 738.317730] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.317730] env[62558]: self.force_reraise() [ 738.317730] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.317730] env[62558]: raise self.value [ 738.317730] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.317730] env[62558]: updated_port = self._update_port( [ 738.317730] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.317730] env[62558]: _ensure_no_port_binding_failure(port) [ 738.317730] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.317730] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.318837] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 738.318837] env[62558]: Removing descriptor: 15 [ 738.512487] env[62558]: DEBUG nova.scheduler.client.report [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.555686] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.588309] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.588309] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.588309] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.588460] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.589580] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.589580] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.590102] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.590653] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.590750] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.593097] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.593097] env[62558]: DEBUG nova.virt.hardware [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.593097] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0dd48b-a20e-4659-bd33-36cb9fcfd5d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.601112] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c291578-f4e7-4f4f-affd-c6992e34255b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.614937] env[62558]: ERROR nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Traceback (most recent call last): [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] yield resources [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.driver.spawn(context, instance, image_meta, [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] vm_ref = self.build_virtual_machine(instance, [ 738.614937] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] for vif in network_info: [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return self._sync_wrapper(fn, *args, **kwargs) [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.wait() [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self[:] = self._gt.wait() [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return self._exit_event.wait() [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.615541] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] current.throw(*self._exc) [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] result = function(*args, **kwargs) [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return func(*args, **kwargs) [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise e [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] nwinfo = self.network_api.allocate_for_instance( [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] created_port_ids = self._update_ports_for_instance( [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] with excutils.save_and_reraise_exception(): [ 738.616218] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.force_reraise() [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise self.value [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] updated_port = self._update_port( [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] _ensure_no_port_binding_failure(port) [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise exception.PortBindingFailed(port_id=port['id']) [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 738.617392] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] [ 738.617392] env[62558]: INFO nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Terminating instance [ 738.617953] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquiring lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.679169] env[62558]: DEBUG nova.network.neutron [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.808447] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.809380] env[62558]: DEBUG nova.network.neutron [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.019482] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.019482] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.031468] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.236s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.031468] env[62558]: INFO nova.compute.claims [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.312905] env[62558]: DEBUG oslo_concurrency.lockutils [req-16e5b797-e84c-4eb8-bf5d-628423298ae1 req-f9c18a2e-b025-4a80-8281-51a60e361a45 service nova] Releasing lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.313381] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquired lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.313542] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.532555] env[62558]: DEBUG nova.compute.utils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.534010] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.536936] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 739.578506] env[62558]: DEBUG nova.policy [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aed5a5eee35e460988fd042c5aad3e7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ef05869d7cd41bb80724770fa7a20a2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.848235] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.886169] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Successfully created port: bafe92d7-5bf8-44bf-b45b-8be12ac9aa82 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.020343] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.037776] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.182542] env[62558]: DEBUG nova.compute.manager [req-0841e7a6-2f26-4a0c-9240-1b31fe58fcc8 req-639fa810-de74-4ec6-8a1f-78cae21acfc0 service nova] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Received event network-vif-deleted-488fd776-f368-4d06-abf6-03c8a8880f86 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.456859] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9c6432-3566-4eae-9d90-92ad328df6f4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.465161] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c841c6-fa65-4eb4-8275-440c39453da1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.501613] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53f3be3-1acf-4c73-b354-968f608b3908 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.510243] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852de9c7-c84d-49b3-ba7c-c55c59a1087e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.522639] env[62558]: DEBUG nova.compute.provider_tree [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.526569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Releasing lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.529024] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.529024] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.529024] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-655ec00e-c107-4184-93a5-56c41bbd6aab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.535899] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e69008-bf21-4370-a176-7a4cd26f4064 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.561167] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3e035ee1-a63a-46a6-84ca-4980e1185e03 could not be found. [ 740.561167] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.561167] env[62558]: INFO nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Took 0.03 seconds to destroy the instance on the hypervisor. [ 740.561413] env[62558]: DEBUG oslo.service.loopingcall [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.561413] env[62558]: DEBUG nova.compute.manager [-] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.561731] env[62558]: DEBUG nova.network.neutron [-] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.583315] env[62558]: DEBUG nova.network.neutron [-] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.908976] env[62558]: ERROR nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 740.908976] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.908976] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.908976] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.908976] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.908976] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.908976] env[62558]: ERROR nova.compute.manager raise self.value [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.908976] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.908976] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.908976] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.909525] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.909525] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.909525] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 740.909525] env[62558]: ERROR nova.compute.manager [ 740.909525] env[62558]: Traceback (most recent call last): [ 740.909525] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.909525] env[62558]: listener.cb(fileno) [ 740.909525] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.909525] env[62558]: result = function(*args, **kwargs) [ 740.909525] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.909525] env[62558]: return func(*args, **kwargs) [ 740.909525] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.909525] env[62558]: raise e [ 740.909525] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.909525] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 740.909525] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.909525] env[62558]: created_port_ids = self._update_ports_for_instance( [ 740.909525] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.909525] env[62558]: with excutils.save_and_reraise_exception(): [ 740.909525] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.909525] env[62558]: self.force_reraise() [ 740.909525] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.909525] env[62558]: raise self.value [ 740.909525] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.909525] env[62558]: updated_port = self._update_port( [ 740.909525] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.909525] env[62558]: _ensure_no_port_binding_failure(port) [ 740.909525] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.909525] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.910376] env[62558]: nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 740.910376] env[62558]: Removing descriptor: 15 [ 741.028025] env[62558]: DEBUG nova.scheduler.client.report [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.054017] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.084783] env[62558]: DEBUG nova.network.neutron [-] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.087867] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.088381] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.088620] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.088851] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.089102] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.089304] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.089567] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.089767] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.089970] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.090202] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.090439] env[62558]: DEBUG nova.virt.hardware [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.091340] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6100cf67-ad5b-40e8-b017-2a12409ef8ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.100939] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5262542-7123-440a-9c71-1e27638e6d4a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.118501] env[62558]: ERROR nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Traceback (most recent call last): [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] yield resources [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.driver.spawn(context, instance, image_meta, [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] vm_ref = self.build_virtual_machine(instance, [ 741.118501] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] for vif in network_info: [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return self._sync_wrapper(fn, *args, **kwargs) [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.wait() [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self[:] = self._gt.wait() [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return self._exit_event.wait() [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.118932] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] current.throw(*self._exc) [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] result = function(*args, **kwargs) [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return func(*args, **kwargs) [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise e [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] nwinfo = self.network_api.allocate_for_instance( [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] created_port_ids = self._update_ports_for_instance( [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] with excutils.save_and_reraise_exception(): [ 741.119328] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.force_reraise() [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise self.value [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] updated_port = self._update_port( [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] _ensure_no_port_binding_failure(port) [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise exception.PortBindingFailed(port_id=port['id']) [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 741.120165] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] [ 741.120165] env[62558]: INFO nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Terminating instance [ 741.121831] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquiring lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.121994] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquired lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.122223] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.531465] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.532041] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.534547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.761s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.535938] env[62558]: INFO nova.compute.claims [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.594970] env[62558]: INFO nova.compute.manager [-] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Took 1.03 seconds to deallocate network for instance. [ 741.598829] env[62558]: DEBUG nova.compute.claims [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.598982] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.639016] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.718352] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.869731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.870019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.040181] env[62558]: DEBUG nova.compute.utils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.046675] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.046675] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.091923] env[62558]: DEBUG nova.policy [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd44eed094d3d4e9f92bd0c99e6566986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7220b5635a34080aeaa7420b92990ad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.207158] env[62558]: DEBUG nova.compute.manager [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Received event network-changed-bafe92d7-5bf8-44bf-b45b-8be12ac9aa82 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.207478] env[62558]: DEBUG nova.compute.manager [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Refreshing instance network info cache due to event network-changed-bafe92d7-5bf8-44bf-b45b-8be12ac9aa82. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 742.208140] env[62558]: DEBUG oslo_concurrency.lockutils [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] Acquiring lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.223324] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Releasing lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.224134] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.224134] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 742.224927] env[62558]: DEBUG oslo_concurrency.lockutils [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] Acquired lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.225162] env[62558]: DEBUG nova.network.neutron [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Refreshing network info cache for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 742.226376] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f74dbc39-6385-44d5-82d4-3398602c4685 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.238699] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b7d516-5ab8-45a3-866a-e3d77a56a2ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.261180] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d526aa96-4007-45f8-8eee-9030bdb9f402 could not be found. [ 742.261405] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 742.261589] env[62558]: INFO nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Took 0.04 seconds to destroy the instance on the hypervisor. [ 742.261843] env[62558]: DEBUG oslo.service.loopingcall [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.262065] env[62558]: DEBUG nova.compute.manager [-] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.262157] env[62558]: DEBUG nova.network.neutron [-] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.278460] env[62558]: DEBUG nova.network.neutron [-] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.279814] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.280032] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.314024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.314024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.335208] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.335417] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.357655] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Successfully created port: d634a7b0-51b7-466d-8e64-2f75affbe7f9 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.544676] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.760900] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Successfully created port: 6ff18bb4-40e5-49e1-886b-aaf3d9883ca4 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.765233] env[62558]: DEBUG nova.network.neutron [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.782040] env[62558]: DEBUG nova.network.neutron [-] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.891852] env[62558]: DEBUG nova.network.neutron [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.960970] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbef9a7f-9cbd-4293-a623-437c50b83ed5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.968789] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ec0b3c-12b1-4969-8aa5-fb2fdf90c06f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.999120] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1272fe7-489d-4aa9-9e3e-91afc3c0cedc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.005908] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ca50aa-30c9-456b-b6d6-76d8df1af954 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.024289] env[62558]: DEBUG nova.compute.provider_tree [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.038459] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Successfully created port: b02100d5-d63a-4708-9731-2b90589e32d3 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.285904] env[62558]: INFO nova.compute.manager [-] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Took 1.02 seconds to deallocate network for instance. [ 743.288323] env[62558]: DEBUG nova.compute.claims [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.288499] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.394207] env[62558]: DEBUG oslo_concurrency.lockutils [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] Releasing lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.394505] env[62558]: DEBUG nova.compute.manager [req-e9782e14-c012-48fd-8a20-a5a84c5c76c1 req-559c36b7-c351-455a-bf5f-6d85227a5ff9 service nova] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Received event network-vif-deleted-bafe92d7-5bf8-44bf-b45b-8be12ac9aa82 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.528080] env[62558]: DEBUG nova.scheduler.client.report [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.557289] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.588940] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.589215] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.589448] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.589634] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.589777] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.589919] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.590201] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.590377] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.590598] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.590788] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.590960] env[62558]: DEBUG nova.virt.hardware [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.591862] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd4ab8d-3768-462b-93fc-0e80e04eab61 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.599376] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37031bb-6049-43b0-8ec2-d37c607314e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.837414] env[62558]: ERROR nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 743.837414] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.837414] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.837414] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.837414] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.837414] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.837414] env[62558]: ERROR nova.compute.manager raise self.value [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.837414] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.837414] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.837414] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.838022] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.838022] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.838022] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 743.838022] env[62558]: ERROR nova.compute.manager [ 743.838022] env[62558]: Traceback (most recent call last): [ 743.838022] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.838022] env[62558]: listener.cb(fileno) [ 743.838022] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.838022] env[62558]: result = function(*args, **kwargs) [ 743.838022] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.838022] env[62558]: return func(*args, **kwargs) [ 743.838022] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.838022] env[62558]: raise e [ 743.838022] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.838022] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 743.838022] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.838022] env[62558]: created_port_ids = self._update_ports_for_instance( [ 743.838022] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.838022] env[62558]: with excutils.save_and_reraise_exception(): [ 743.838022] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.838022] env[62558]: self.force_reraise() [ 743.838022] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.838022] env[62558]: raise self.value [ 743.838022] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.838022] env[62558]: updated_port = self._update_port( [ 743.838022] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.838022] env[62558]: _ensure_no_port_binding_failure(port) [ 743.838022] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.838022] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.839015] env[62558]: nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 743.839015] env[62558]: Removing descriptor: 15 [ 743.839015] env[62558]: ERROR nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Traceback (most recent call last): [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] yield resources [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.driver.spawn(context, instance, image_meta, [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.839015] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] vm_ref = self.build_virtual_machine(instance, [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] for vif in network_info: [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self._sync_wrapper(fn, *args, **kwargs) [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.wait() [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self[:] = self._gt.wait() [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self._exit_event.wait() [ 743.839443] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] result = hub.switch() [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self.greenlet.switch() [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] result = function(*args, **kwargs) [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return func(*args, **kwargs) [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise e [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] nwinfo = self.network_api.allocate_for_instance( [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.839883] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] created_port_ids = self._update_ports_for_instance( [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] with excutils.save_and_reraise_exception(): [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.force_reraise() [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise self.value [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] updated_port = self._update_port( [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] _ensure_no_port_binding_failure(port) [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.840340] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise exception.PortBindingFailed(port_id=port['id']) [ 743.840753] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 743.840753] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] [ 743.840753] env[62558]: INFO nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Terminating instance [ 743.840753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.840753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.840753] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.034053] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.034167] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.036599] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.103s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.262766] env[62558]: DEBUG nova.compute.manager [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Received event network-changed-d634a7b0-51b7-466d-8e64-2f75affbe7f9 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.262976] env[62558]: DEBUG nova.compute.manager [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Refreshing instance network info cache due to event network-changed-d634a7b0-51b7-466d-8e64-2f75affbe7f9. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.263131] env[62558]: DEBUG oslo_concurrency.lockutils [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] Acquiring lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.384792] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.474219] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.541030] env[62558]: DEBUG nova.compute.utils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.545612] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.546409] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.585905] env[62558]: DEBUG nova.policy [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b177e8a629d2494aaf0497b520408909', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50cccd70acb04a4aa4691260e3437a94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.833966] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Successfully created port: a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.900555] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f493899-cdab-4256-b16b-f98672a97c39 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.907009] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4083a6d-9108-47c8-9443-9c85c6d966a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.937952] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f559f995-06a8-4170-8117-26e99d216654 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.945263] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4485ef-7f8b-49c8-9592-6c8508178892 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.958135] env[62558]: DEBUG nova.compute.provider_tree [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.976932] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Releasing lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.977373] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 744.977563] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.977862] env[62558]: DEBUG oslo_concurrency.lockutils [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] Acquired lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.978038] env[62558]: DEBUG nova.network.neutron [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Refreshing network info cache for port d634a7b0-51b7-466d-8e64-2f75affbe7f9 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.978975] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebfecd9a-7783-493f-a719-a120d955bd77 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.987839] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05007e82-6b46-4e33-9699-bbe51078945d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.008949] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d92dca09-0a42-4103-ac92-25b6b0fa3c32 could not be found. [ 745.009264] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.009416] env[62558]: INFO nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.009663] env[62558]: DEBUG oslo.service.loopingcall [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.009866] env[62558]: DEBUG nova.compute.manager [-] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.009961] env[62558]: DEBUG nova.network.neutron [-] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.046128] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.068174] env[62558]: DEBUG nova.network.neutron [-] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.460887] env[62558]: DEBUG nova.scheduler.client.report [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.552080] env[62558]: DEBUG nova.network.neutron [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.675943] env[62558]: ERROR nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 745.675943] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.675943] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.675943] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.675943] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.675943] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.675943] env[62558]: ERROR nova.compute.manager raise self.value [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.675943] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 745.675943] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.675943] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 745.677014] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.677014] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 745.677014] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 745.677014] env[62558]: ERROR nova.compute.manager [ 745.677014] env[62558]: Traceback (most recent call last): [ 745.677014] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 745.677014] env[62558]: listener.cb(fileno) [ 745.677014] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.677014] env[62558]: result = function(*args, **kwargs) [ 745.677014] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.677014] env[62558]: return func(*args, **kwargs) [ 745.677014] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.677014] env[62558]: raise e [ 745.677014] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.677014] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 745.677014] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.677014] env[62558]: created_port_ids = self._update_ports_for_instance( [ 745.677014] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.677014] env[62558]: with excutils.save_and_reraise_exception(): [ 745.677014] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.677014] env[62558]: self.force_reraise() [ 745.677014] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.677014] env[62558]: raise self.value [ 745.677014] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.677014] env[62558]: updated_port = self._update_port( [ 745.677014] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.677014] env[62558]: _ensure_no_port_binding_failure(port) [ 745.677014] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.677014] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 745.677932] env[62558]: nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 745.677932] env[62558]: Removing descriptor: 15 [ 745.752142] env[62558]: DEBUG nova.network.neutron [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.966208] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.966911] env[62558]: ERROR nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Traceback (most recent call last): [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.driver.spawn(context, instance, image_meta, [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] vm_ref = self.build_virtual_machine(instance, [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.966911] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] for vif in network_info: [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self._sync_wrapper(fn, *args, **kwargs) [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.wait() [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self[:] = self._gt.wait() [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self._exit_event.wait() [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] result = hub.switch() [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.967352] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return self.greenlet.switch() [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] result = function(*args, **kwargs) [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] return func(*args, **kwargs) [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise e [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] nwinfo = self.network_api.allocate_for_instance( [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] created_port_ids = self._update_ports_for_instance( [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] with excutils.save_and_reraise_exception(): [ 745.967761] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] self.force_reraise() [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise self.value [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] updated_port = self._update_port( [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] _ensure_no_port_binding_failure(port) [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] raise exception.PortBindingFailed(port_id=port['id']) [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] nova.exception.PortBindingFailed: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. [ 745.968194] env[62558]: ERROR nova.compute.manager [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] [ 745.968549] env[62558]: DEBUG nova.compute.utils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.969373] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.741s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.970848] env[62558]: INFO nova.compute.claims [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.973635] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Build of instance 151768fd-c2c8-491f-93b5-bbc53bd69e2c was re-scheduled: Binding failed for port 06567788-b01a-48a6-a6c0-27fa9897a25f, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.974297] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.974297] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.974431] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquired lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.974526] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.058925] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.083724] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.083970] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.084138] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.084317] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.084460] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.084604] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.084801] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.084972] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.085153] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.085312] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.085478] env[62558]: DEBUG nova.virt.hardware [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.086332] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbc4bdf-5e15-4722-a517-820dd15fc256 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.094515] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27eae807-8309-4dfa-899c-198bb24f9c6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.107926] env[62558]: ERROR nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Traceback (most recent call last): [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] yield resources [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.driver.spawn(context, instance, image_meta, [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] vm_ref = self.build_virtual_machine(instance, [ 746.107926] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] for vif in network_info: [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return self._sync_wrapper(fn, *args, **kwargs) [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.wait() [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self[:] = self._gt.wait() [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return self._exit_event.wait() [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.108410] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] current.throw(*self._exc) [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] result = function(*args, **kwargs) [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return func(*args, **kwargs) [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise e [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] nwinfo = self.network_api.allocate_for_instance( [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] created_port_ids = self._update_ports_for_instance( [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] with excutils.save_and_reraise_exception(): [ 746.108833] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.force_reraise() [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise self.value [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] updated_port = self._update_port( [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] _ensure_no_port_binding_failure(port) [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise exception.PortBindingFailed(port_id=port['id']) [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 746.109273] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] [ 746.109273] env[62558]: INFO nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Terminating instance [ 746.110243] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.110394] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquired lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.110555] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.254781] env[62558]: DEBUG oslo_concurrency.lockutils [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] Releasing lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.254969] env[62558]: DEBUG nova.compute.manager [req-453c48fe-33fb-4fb3-9197-9ede66c5061e req-5ef409f5-5d1b-4327-bd03-8bfc1ee062ae service nova] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Received event network-vif-deleted-d634a7b0-51b7-466d-8e64-2f75affbe7f9 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.287866] env[62558]: DEBUG nova.compute.manager [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Received event network-changed-a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.288036] env[62558]: DEBUG nova.compute.manager [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Refreshing instance network info cache due to event network-changed-a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.288244] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] Acquiring lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.381614] env[62558]: DEBUG nova.network.neutron [-] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.493572] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.544038] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.636176] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.724857] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.884069] env[62558]: INFO nova.compute.manager [-] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Took 1.87 seconds to deallocate network for instance. [ 746.886294] env[62558]: DEBUG nova.compute.claims [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 746.886465] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.045998] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Releasing lock "refresh_cache-151768fd-c2c8-491f-93b5-bbc53bd69e2c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.046267] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 747.046434] env[62558]: DEBUG nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.046601] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.067164] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.229367] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Releasing lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.229795] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.229986] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.230518] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] Acquired lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.231115] env[62558]: DEBUG nova.network.neutron [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Refreshing network info cache for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.232840] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60f074c1-b495-4d70-90a9-37c88dc1d966 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.244162] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf9fdae-6e5b-4a8b-bcfe-f9429c31ee7d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.266453] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0 could not be found. [ 747.266543] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.266697] env[62558]: INFO nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 747.267168] env[62558]: DEBUG oslo.service.loopingcall [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.269185] env[62558]: DEBUG nova.compute.manager [-] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.269298] env[62558]: DEBUG nova.network.neutron [-] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.292193] env[62558]: DEBUG nova.network.neutron [-] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.337987] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b1ad9b-8593-4554-84b2-5a53c41dd117 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.345334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9096164-9ac4-441f-bf4b-7c2452d3c1bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.374096] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eac924a-8505-418b-9d46-a4a63b958625 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.381049] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84231fe1-2b45-441c-a260-4bf9ca293d97 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.393808] env[62558]: DEBUG nova.compute.provider_tree [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.569850] env[62558]: DEBUG nova.network.neutron [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.755195] env[62558]: DEBUG nova.network.neutron [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.794487] env[62558]: DEBUG nova.network.neutron [-] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.861215] env[62558]: DEBUG nova.network.neutron [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.896695] env[62558]: DEBUG nova.scheduler.client.report [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.074685] env[62558]: INFO nova.compute.manager [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: 151768fd-c2c8-491f-93b5-bbc53bd69e2c] Took 1.03 seconds to deallocate network for instance. [ 748.297051] env[62558]: INFO nova.compute.manager [-] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Took 1.03 seconds to deallocate network for instance. [ 748.299218] env[62558]: DEBUG nova.compute.claims [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.299454] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.363711] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] Releasing lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.363968] env[62558]: DEBUG nova.compute.manager [req-b1fd3f89-475b-45ee-8872-f018449be616 req-6783f4fe-e67f-4189-a261-63c2372886b0 service nova] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Received event network-vif-deleted-a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.401414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.401969] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.404498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.223s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.405897] env[62558]: INFO nova.compute.claims [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.914250] env[62558]: DEBUG nova.compute.utils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.915123] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.915294] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.950480] env[62558]: DEBUG nova.policy [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbd32f833a88422f850c86d764de3736', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81ebf6dfdbcd46098ad0e978899cd27b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.105650] env[62558]: INFO nova.scheduler.client.report [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Deleted allocations for instance 151768fd-c2c8-491f-93b5-bbc53bd69e2c [ 749.274607] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Successfully created port: 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.418390] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.612645] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e5e7504b-9be5-4b40-919e-9332f8ab8c73 tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "151768fd-c2c8-491f-93b5-bbc53bd69e2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.499s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.784436] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51da7ffd-0473-4f9a-8fc9-8796c49abd00 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.793243] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d82ee90-c089-4a01-9804-bf4524344bfe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.836016] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349c8c19-2427-4d57-8407-c07add9fb2ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.844395] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba37b2a-a7e2-4010-92b1-8463a8b12d6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.857631] env[62558]: DEBUG nova.compute.provider_tree [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.115480] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.232468] env[62558]: DEBUG nova.compute.manager [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Received event network-changed-1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.232673] env[62558]: DEBUG nova.compute.manager [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Refreshing instance network info cache due to event network-changed-1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.232888] env[62558]: DEBUG oslo_concurrency.lockutils [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] Acquiring lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.233038] env[62558]: DEBUG oslo_concurrency.lockutils [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] Acquired lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.233197] env[62558]: DEBUG nova.network.neutron [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Refreshing network info cache for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.362499] env[62558]: DEBUG nova.scheduler.client.report [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.432557] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.435858] env[62558]: ERROR nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 750.435858] env[62558]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.435858] env[62558]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.435858] env[62558]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.435858] env[62558]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.435858] env[62558]: ERROR nova.compute.manager self.force_reraise() [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.435858] env[62558]: ERROR nova.compute.manager raise self.value [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.435858] env[62558]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.435858] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.435858] env[62558]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.436581] env[62558]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.436581] env[62558]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.436581] env[62558]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 750.436581] env[62558]: ERROR nova.compute.manager [ 750.436581] env[62558]: Traceback (most recent call last): [ 750.436581] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.436581] env[62558]: listener.cb(fileno) [ 750.436581] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.436581] env[62558]: result = function(*args, **kwargs) [ 750.436581] env[62558]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.436581] env[62558]: return func(*args, **kwargs) [ 750.436581] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.436581] env[62558]: raise e [ 750.436581] env[62558]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.436581] env[62558]: nwinfo = self.network_api.allocate_for_instance( [ 750.436581] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.436581] env[62558]: created_port_ids = self._update_ports_for_instance( [ 750.436581] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.436581] env[62558]: with excutils.save_and_reraise_exception(): [ 750.436581] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.436581] env[62558]: self.force_reraise() [ 750.436581] env[62558]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.436581] env[62558]: raise self.value [ 750.436581] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.436581] env[62558]: updated_port = self._update_port( [ 750.436581] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.436581] env[62558]: _ensure_no_port_binding_failure(port) [ 750.436581] env[62558]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.436581] env[62558]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.438141] env[62558]: nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 750.438141] env[62558]: Removing descriptor: 22 [ 750.458447] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.458693] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.458868] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.459058] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.459205] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.459368] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.459654] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.459847] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.460047] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.460218] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.460410] env[62558]: DEBUG nova.virt.hardware [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.461349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584f6d4e-8df8-4ad3-bf41-07a75fadeb86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.469811] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bada33ce-5eb5-4491-b152-42e1e37f8f5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.484191] env[62558]: ERROR nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Traceback (most recent call last): [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] yield resources [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.driver.spawn(context, instance, image_meta, [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] vm_ref = self.build_virtual_machine(instance, [ 750.484191] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] for vif in network_info: [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return self._sync_wrapper(fn, *args, **kwargs) [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.wait() [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self[:] = self._gt.wait() [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return self._exit_event.wait() [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.484580] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] current.throw(*self._exc) [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] result = function(*args, **kwargs) [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return func(*args, **kwargs) [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise e [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] nwinfo = self.network_api.allocate_for_instance( [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] created_port_ids = self._update_ports_for_instance( [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] with excutils.save_and_reraise_exception(): [ 750.484985] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.force_reraise() [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise self.value [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] updated_port = self._update_port( [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] _ensure_no_port_binding_failure(port) [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise exception.PortBindingFailed(port_id=port['id']) [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 750.485489] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] [ 750.485489] env[62558]: INFO nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Terminating instance [ 750.486385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.635775] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.750819] env[62558]: DEBUG nova.network.neutron [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.846745] env[62558]: DEBUG nova.network.neutron [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.866755] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.867793] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.871277] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.611s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.873926] env[62558]: INFO nova.compute.claims [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.350043] env[62558]: DEBUG oslo_concurrency.lockutils [req-4a72b493-a616-471e-8342-f23ad825e27e req-889cb735-2c3d-4c22-baa5-d7a0cefe3cff service nova] Releasing lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.350043] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquired lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.350330] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.377669] env[62558]: DEBUG nova.compute.utils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.380728] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.380790] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.431332] env[62558]: DEBUG nova.policy [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0875092c7b544128e56a2e924838d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e51dcdbfe744ddf909800d2cdc1b98c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.743158] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Successfully created port: 99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.880414] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.888101] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.047499] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.260181] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425bb46a-8387-4b36-80d1-ae7371ffedfe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.267712] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d986ea5e-c4a1-45e6-af30-79d2bbc06048 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.298992] env[62558]: DEBUG nova.compute.manager [req-99e712ae-5ca3-4f04-98d9-4c5850699aaf req-3071dde2-96c0-45f7-b07b-2b5cf075649c service nova] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Received event network-vif-deleted-1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.299893] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaf2b71-3eab-4892-b4e1-ed4ae8318246 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.307830] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a57049f-79f0-402c-8842-ab30bf31c13b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.323795] env[62558]: DEBUG nova.compute.provider_tree [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.550410] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Releasing lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.550839] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.551037] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.551347] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ed0e633-4bb7-4d9b-9a51-ce0d6394f8ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.560926] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccbaf41e-c30b-4d5f-9a01-d0766aab6cc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.581365] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d65fbfe3-b1e9-48d9-88ba-209fdc264283 could not be found. [ 752.581574] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.581749] env[62558]: INFO nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.581984] env[62558]: DEBUG oslo.service.loopingcall [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.582208] env[62558]: DEBUG nova.compute.manager [-] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.582299] env[62558]: DEBUG nova.network.neutron [-] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.735404] env[62558]: DEBUG nova.network.neutron [-] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.827762] env[62558]: DEBUG nova.scheduler.client.report [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.902089] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.929018] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.929018] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.929018] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.929402] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.929402] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.929402] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.930039] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.931022] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.931106] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.931340] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.931537] env[62558]: DEBUG nova.virt.hardware [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.932564] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a52fd4-0cad-444e-a0c3-2ed05c5a7a63 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.940719] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1fd916-1b90-4478-a331-663086270ea8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.238192] env[62558]: DEBUG nova.network.neutron [-] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.332168] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.332916] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 753.335551] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.878s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.335727] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.335878] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 753.336173] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.248s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.339489] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c6355f-8045-4666-985a-48016bf0c284 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.348173] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a83b763-6682-4a85-9096-8c0e48cdf1e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.364154] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2dfc2a-d498-40bc-bac6-f64b3259c7db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.371366] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80f2c75-fb28-4182-9473-e9de4572a2b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.403918] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181465MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 753.404104] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.608391] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Successfully updated port: 99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 753.740539] env[62558]: INFO nova.compute.manager [-] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Took 1.16 seconds to deallocate network for instance. [ 753.743372] env[62558]: DEBUG nova.compute.claims [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Aborting claim: {{(pid=62558) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.743515] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.840734] env[62558]: DEBUG nova.compute.utils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.842197] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 753.842363] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 753.902758] env[62558]: DEBUG nova.policy [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0875092c7b544128e56a2e924838d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e51dcdbfe744ddf909800d2cdc1b98c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.113278] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.113429] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.113567] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.194462] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f07620-8238-4725-8365-7687e8cf7def {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.203803] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afab027-f498-4601-b7a8-d1fbc21e86f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.233790] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0802350d-754c-43ca-9869-479fa9460aa4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.244174] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b9aa67-803d-45eb-a151-bb1794df99db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.245478] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Successfully created port: 9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.256944] env[62558]: DEBUG nova.compute.provider_tree [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.321310] env[62558]: DEBUG nova.compute.manager [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Received event network-vif-plugged-99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.321534] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Acquiring lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.321736] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.321987] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.322173] env[62558]: DEBUG nova.compute.manager [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] No waiting events found dispatching network-vif-plugged-99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 754.322372] env[62558]: WARNING nova.compute.manager [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Received unexpected event network-vif-plugged-99b91852-f90d-4a37-9558-042993918fd5 for instance with vm_state building and task_state spawning. [ 754.322554] env[62558]: DEBUG nova.compute.manager [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Received event network-changed-99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.322724] env[62558]: DEBUG nova.compute.manager [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Refreshing instance network info cache due to event network-changed-99b91852-f90d-4a37-9558-042993918fd5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.322888] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Acquiring lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.344973] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 754.651855] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.759971] env[62558]: DEBUG nova.scheduler.client.report [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.806792] env[62558]: DEBUG nova.network.neutron [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updating instance_info_cache with network_info: [{"id": "99b91852-f90d-4a37-9558-042993918fd5", "address": "fa:16:3e:7b:40:aa", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99b91852-f9", "ovs_interfaceid": "99b91852-f90d-4a37-9558-042993918fd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.264375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.265033] env[62558]: ERROR nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Traceback (most recent call last): [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.driver.spawn(context, instance, image_meta, [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] vm_ref = self.build_virtual_machine(instance, [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.265033] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] for vif in network_info: [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return self._sync_wrapper(fn, *args, **kwargs) [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.wait() [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self[:] = self._gt.wait() [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return self._exit_event.wait() [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] current.throw(*self._exc) [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.265411] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] result = function(*args, **kwargs) [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] return func(*args, **kwargs) [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise e [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] nwinfo = self.network_api.allocate_for_instance( [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] created_port_ids = self._update_ports_for_instance( [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] with excutils.save_and_reraise_exception(): [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] self.force_reraise() [ 755.265796] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise self.value [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] updated_port = self._update_port( [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] _ensure_no_port_binding_failure(port) [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] raise exception.PortBindingFailed(port_id=port['id']) [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] nova.exception.PortBindingFailed: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. [ 755.266257] env[62558]: ERROR nova.compute.manager [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] [ 755.266257] env[62558]: DEBUG nova.compute.utils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.266958] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.459s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.268677] env[62558]: INFO nova.compute.claims [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.271204] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Build of instance f3128303-9b2c-4f14-9ea0-093f55ecb063 was re-scheduled: Binding failed for port c027a1a0-5707-4c6b-9967-6c4973090fde, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 755.271619] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 755.271840] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquiring lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.271982] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Acquired lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.272148] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.309043] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.309298] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance network_info: |[{"id": "99b91852-f90d-4a37-9558-042993918fd5", "address": "fa:16:3e:7b:40:aa", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99b91852-f9", "ovs_interfaceid": "99b91852-f90d-4a37-9558-042993918fd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 755.309597] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Acquired lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.309781] env[62558]: DEBUG nova.network.neutron [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Refreshing network info cache for port 99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.310909] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:40:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed3ffc1d-9f86-4029-857e-6cd1d383edbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99b91852-f90d-4a37-9558-042993918fd5', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.318502] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Creating folder: Project (0e51dcdbfe744ddf909800d2cdc1b98c). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.319353] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e50a5b10-20cb-46b1-8765-ac55ca217211 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.330264] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Created folder: Project (0e51dcdbfe744ddf909800d2cdc1b98c) in parent group-v272451. [ 755.330443] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Creating folder: Instances. Parent ref: group-v272472. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.330681] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43d5159d-5034-4d95-a684-91fe422b6b8b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.340429] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Created folder: Instances in parent group-v272472. [ 755.340684] env[62558]: DEBUG oslo.service.loopingcall [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.340871] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.341072] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aee61627-5559-4efb-911b-ed71e499744e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.355498] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.362884] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.362884] env[62558]: value = "task-1266557" [ 755.362884] env[62558]: _type = "Task" [ 755.362884] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.370330] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266557, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.379541] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.379757] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.379910] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.380103] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.380252] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.380397] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.380590] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.380745] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.380902] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.381069] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.381241] env[62558]: DEBUG nova.virt.hardware [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.381981] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c7c99d-d285-4aaf-9195-52c9d8fd97aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.388585] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e435efce-13aa-48cd-baa0-bc42b8966669 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.792342] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.872950] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266557, 'name': CreateVM_Task, 'duration_secs': 0.368302} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.873175] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.875688] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.885579] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.885828] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.886236] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.886544] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-000f2a8a-faf8-4161-8fb5-19c9644d7dc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.897590] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 755.897590] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52128bad-3f11-ca73-8673-78ec86c009a3" [ 755.897590] env[62558]: _type = "Task" [ 755.897590] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.898601] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Successfully updated port: 9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 755.908686] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52128bad-3f11-ca73-8673-78ec86c009a3, 'name': SearchDatastore_Task, 'duration_secs': 0.010193} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.911274] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.911522] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.911749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.911919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.912126] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.913017] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c80176b7-a9ab-4ba0-aae5-c78b216a016b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.921435] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.921636] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.922606] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d73cbf8-8b1c-4209-83b4-c79e6a6cff08 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.927353] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 755.927353] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520edde7-0cdc-8870-b1ec-d70a55ead37a" [ 755.927353] env[62558]: _type = "Task" [ 755.927353] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.939946] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520edde7-0cdc-8870-b1ec-d70a55ead37a, 'name': SearchDatastore_Task, 'duration_secs': 0.008204} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.947964] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-337046f4-09d4-401a-8558-11da93c38585 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.955353] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 755.955353] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ab1d6-4177-f036-a680-79c03e94a0b1" [ 755.955353] env[62558]: _type = "Task" [ 755.955353] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.960603] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ab1d6-4177-f036-a680-79c03e94a0b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.084543] env[62558]: DEBUG nova.network.neutron [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updated VIF entry in instance network info cache for port 99b91852-f90d-4a37-9558-042993918fd5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 756.084911] env[62558]: DEBUG nova.network.neutron [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updating instance_info_cache with network_info: [{"id": "99b91852-f90d-4a37-9558-042993918fd5", "address": "fa:16:3e:7b:40:aa", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99b91852-f9", "ovs_interfaceid": "99b91852-f90d-4a37-9558-042993918fd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.347487] env[62558]: DEBUG nova.compute.manager [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Received event network-vif-plugged-9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.347735] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Acquiring lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.347936] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.348112] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.348285] env[62558]: DEBUG nova.compute.manager [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] No waiting events found dispatching network-vif-plugged-9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 756.348444] env[62558]: WARNING nova.compute.manager [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Received unexpected event network-vif-plugged-9245beb2-0701-4744-b916-d8291b7b8db0 for instance with vm_state building and task_state spawning. [ 756.348597] env[62558]: DEBUG nova.compute.manager [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Received event network-changed-9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.348741] env[62558]: DEBUG nova.compute.manager [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Refreshing instance network info cache due to event network-changed-9245beb2-0701-4744-b916-d8291b7b8db0. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 756.348930] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Acquiring lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.349046] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Acquired lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.349198] env[62558]: DEBUG nova.network.neutron [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Refreshing network info cache for port 9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 756.378869] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Releasing lock "refresh_cache-f3128303-9b2c-4f14-9ea0-093f55ecb063" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.379124] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.379290] env[62558]: DEBUG nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.379564] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.395896] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.400869] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.462010] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ab1d6-4177-f036-a680-79c03e94a0b1, 'name': SearchDatastore_Task, 'duration_secs': 0.008351} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.464217] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.464477] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 867c0fb3-d4f6-4945-a5cc-6d93efcadb79/867c0fb3-d4f6-4945-a5cc-6d93efcadb79.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.465024] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c0b932f-9011-4db8-9c34-875d2d366f33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.471908] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 756.471908] env[62558]: value = "task-1266558" [ 756.471908] env[62558]: _type = "Task" [ 756.471908] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.481961] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.587792] env[62558]: DEBUG oslo_concurrency.lockutils [req-e3033231-eea6-4b0d-9438-ba06a1cf7dd5 req-e062acfd-5278-4bd0-af73-664706ce423b service nova] Releasing lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.639821] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3203369-13dc-4b65-9b31-88cdeeaa26c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.649508] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c15228-0b17-4d9f-9d74-4bfcae22fa5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.684730] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b73dab6-223e-40e0-9fad-d22f81ebecea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.692786] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fcef4e-3bb1-4e90-851b-05e40b629c80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.707428] env[62558]: DEBUG nova.compute.provider_tree [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.882342] env[62558]: DEBUG nova.network.neutron [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.898127] env[62558]: DEBUG nova.network.neutron [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.958197] env[62558]: DEBUG nova.network.neutron [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.982099] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477857} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.982464] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 867c0fb3-d4f6-4945-a5cc-6d93efcadb79/867c0fb3-d4f6-4945-a5cc-6d93efcadb79.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 756.982576] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.982784] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eed94067-3b40-40bc-8c92-8059528f655e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.988138] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 756.988138] env[62558]: value = "task-1266559" [ 756.988138] env[62558]: _type = "Task" [ 756.988138] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.996218] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.212207] env[62558]: DEBUG nova.scheduler.client.report [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.401503] env[62558]: INFO nova.compute.manager [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] [instance: f3128303-9b2c-4f14-9ea0-093f55ecb063] Took 1.02 seconds to deallocate network for instance. [ 757.460590] env[62558]: DEBUG oslo_concurrency.lockutils [req-6d0392eb-aec1-4a91-90b5-c1f3aaac4fa2 req-5f758146-9771-4ce4-b09b-84a74b66e957 service nova] Releasing lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.460979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.461153] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.497761] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067737} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.497761] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.498479] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a54105-0d6e-4d22-9342-c926911d6928 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.520802] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 867c0fb3-d4f6-4945-a5cc-6d93efcadb79/867c0fb3-d4f6-4945-a5cc-6d93efcadb79.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.521076] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54fde4a3-fa30-446d-bc00-ccbe2b96848f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.539961] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 757.539961] env[62558]: value = "task-1266560" [ 757.539961] env[62558]: _type = "Task" [ 757.539961] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.547416] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266560, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.716097] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.716762] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.719587] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.121s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.988897] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.050057] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266560, 'name': ReconfigVM_Task, 'duration_secs': 0.302195} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.052652] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 867c0fb3-d4f6-4945-a5cc-6d93efcadb79/867c0fb3-d4f6-4945-a5cc-6d93efcadb79.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.053333] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-729cf17b-ed81-453f-9961-d204a6b4a94d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.059451] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 758.059451] env[62558]: value = "task-1266561" [ 758.059451] env[62558]: _type = "Task" [ 758.059451] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.067479] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266561, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.121614] env[62558]: DEBUG nova.network.neutron [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Updating instance_info_cache with network_info: [{"id": "9245beb2-0701-4744-b916-d8291b7b8db0", "address": "fa:16:3e:8f:51:1f", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9245beb2-07", "ovs_interfaceid": "9245beb2-0701-4744-b916-d8291b7b8db0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.224306] env[62558]: DEBUG nova.compute.utils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.225902] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 758.225902] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.285975] env[62558]: DEBUG nova.policy [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0875092c7b544128e56a2e924838d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e51dcdbfe744ddf909800d2cdc1b98c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.446503] env[62558]: INFO nova.scheduler.client.report [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Deleted allocations for instance f3128303-9b2c-4f14-9ea0-093f55ecb063 [ 758.574153] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266561, 'name': Rename_Task, 'duration_secs': 0.151772} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.574153] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.574153] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d96ef79b-2aad-4865-ba3e-7b506854e4e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.577675] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 758.577675] env[62558]: value = "task-1266562" [ 758.577675] env[62558]: _type = "Task" [ 758.577675] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.593737] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.617361] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Successfully created port: bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.623957] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "refresh_cache-6e01c855-ef25-45c9-a73e-5656e8e05cd4" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.624297] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance network_info: |[{"id": "9245beb2-0701-4744-b916-d8291b7b8db0", "address": "fa:16:3e:8f:51:1f", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9245beb2-07", "ovs_interfaceid": "9245beb2-0701-4744-b916-d8291b7b8db0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 758.625518] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:51:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed3ffc1d-9f86-4029-857e-6cd1d383edbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9245beb2-0701-4744-b916-d8291b7b8db0', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.633056] env[62558]: DEBUG oslo.service.loopingcall [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.633197] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.633418] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-176c4fc8-4565-4385-a9ec-2d0e4fdcf75a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.656523] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.656523] env[62558]: value = "task-1266563" [ 758.656523] env[62558]: _type = "Task" [ 758.656523] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.665342] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266563, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.671376] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8afa3f-4292-40ba-b8ad-708294a44b83 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.678396] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208afa45-7c5e-4e6e-b3d2-1b711f0119b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.709082] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b039c639-c677-4a12-b75f-2b5287620d5b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.716929] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ceae5b-a8ec-43a5-844b-c02696d537e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.732024] env[62558]: DEBUG nova.compute.provider_tree [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.732628] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.957891] env[62558]: DEBUG oslo_concurrency.lockutils [None req-466dc0b7-d7af-476f-ac0b-3d23ad4b61cd tempest-ListImageFiltersTestJSON-659361099 tempest-ListImageFiltersTestJSON-659361099-project-member] Lock "f3128303-9b2c-4f14-9ea0-093f55ecb063" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.516s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.090219] env[62558]: DEBUG oslo_vmware.api [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266562, 'name': PowerOnVM_Task, 'duration_secs': 0.502025} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.090471] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.090737] env[62558]: INFO nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Took 6.19 seconds to spawn the instance on the hypervisor. [ 759.090939] env[62558]: DEBUG nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 759.091691] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b35a585-48bd-4023-99a1-4bec36ffc670 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.168274] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266563, 'name': CreateVM_Task, 'duration_secs': 0.30046} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.169077] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 759.169226] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.169290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.169674] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 759.169923] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aa25840-e281-4005-be48-0dabc9932856 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.176835] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 759.176835] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c0db33-6da2-c0c3-d609-54f768e13363" [ 759.176835] env[62558]: _type = "Task" [ 759.176835] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.185104] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c0db33-6da2-c0c3-d609-54f768e13363, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.238080] env[62558]: DEBUG nova.scheduler.client.report [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.460904] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.607601] env[62558]: INFO nova.compute.manager [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Took 27.44 seconds to build instance. [ 759.687173] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c0db33-6da2-c0c3-d609-54f768e13363, 'name': SearchDatastore_Task, 'duration_secs': 0.012598} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.687486] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.687764] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.687994] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.688149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.688325] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.688578] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfb08de9-8099-49dd-85a6-99a91bf34fbe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.696411] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.696582] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 759.697341] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a831484-ce45-4d88-b0ce-e4818609bd76 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.702523] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 759.702523] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5265d129-49fc-fb44-c24c-93b310195616" [ 759.702523] env[62558]: _type = "Task" [ 759.702523] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.709352] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5265d129-49fc-fb44-c24c-93b310195616, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.748345] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.748969] env[62558]: ERROR nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Traceback (most recent call last): [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.driver.spawn(context, instance, image_meta, [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] vm_ref = self.build_virtual_machine(instance, [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.748969] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] for vif in network_info: [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return self._sync_wrapper(fn, *args, **kwargs) [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.wait() [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self[:] = self._gt.wait() [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return self._exit_event.wait() [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] current.throw(*self._exc) [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.749317] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] result = function(*args, **kwargs) [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] return func(*args, **kwargs) [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise e [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] nwinfo = self.network_api.allocate_for_instance( [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] created_port_ids = self._update_ports_for_instance( [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] with excutils.save_and_reraise_exception(): [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] self.force_reraise() [ 759.749735] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise self.value [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] updated_port = self._update_port( [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] _ensure_no_port_binding_failure(port) [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] raise exception.PortBindingFailed(port_id=port['id']) [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] nova.exception.PortBindingFailed: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. [ 759.750308] env[62558]: ERROR nova.compute.manager [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] [ 759.750308] env[62558]: DEBUG nova.compute.utils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.751682] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.753703] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Build of instance 3e035ee1-a63a-46a6-84ca-4980e1185e03 was re-scheduled: Binding failed for port 488fd776-f368-4d06-abf6-03c8a8880f86, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.754329] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.754623] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquiring lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.754777] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Acquired lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.754934] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.757201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.468s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.777731] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.777731] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.780213] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.780213] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.780213] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.780213] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.780213] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.780444] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.780444] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.780444] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.780444] env[62558]: DEBUG nova.virt.hardware [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.780610] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5845320a-b995-4a81-ba27-bc8ac4111b05 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.789077] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f4d0c9-48a3-4267-90d3-9c95ab7f66ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.003713] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.099889] env[62558]: DEBUG nova.compute.manager [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Received event network-vif-plugged-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.099889] env[62558]: DEBUG oslo_concurrency.lockutils [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] Acquiring lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.100467] env[62558]: DEBUG oslo_concurrency.lockutils [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] Lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.100516] env[62558]: DEBUG oslo_concurrency.lockutils [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] Lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.100724] env[62558]: DEBUG nova.compute.manager [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] No waiting events found dispatching network-vif-plugged-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 760.100913] env[62558]: WARNING nova.compute.manager [req-7a873ee1-e5de-4f09-a743-6e1ed6283299 req-9c772f56-482f-4ecc-bf8e-c8844078f2bc service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Received unexpected event network-vif-plugged-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 for instance with vm_state building and task_state spawning. [ 760.109321] env[62558]: DEBUG oslo_concurrency.lockutils [None req-572cacb3-dac0-4e44-8caf-d24e8102942c tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.914s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.193951] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Successfully updated port: bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 760.215095] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5265d129-49fc-fb44-c24c-93b310195616, 'name': SearchDatastore_Task, 'duration_secs': 0.008201} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.215895] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-287eb1f4-b893-42aa-b0a0-2eabdce1b89d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.221138] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 760.221138] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a03ee1-3627-f782-0681-cdb9254082d3" [ 760.221138] env[62558]: _type = "Task" [ 760.221138] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.228224] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a03ee1-3627-f782-0681-cdb9254082d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.279007] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.360081] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.612577] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.671140] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f97169-57ad-44b0-9b12-915ea016418b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.678905] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f196105d-b837-4223-8ed9-23c0366353b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.708302] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.708476] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.708591] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.710281] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0d9750-3243-4292-b872-72c10901c4e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.717988] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707b1657-4fd7-4b33-9a56-0fb495e9d1fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.734112] env[62558]: DEBUG nova.compute.provider_tree [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.738148] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a03ee1-3627-f782-0681-cdb9254082d3, 'name': SearchDatastore_Task, 'duration_secs': 0.01006} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.738564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.738807] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6e01c855-ef25-45c9-a73e-5656e8e05cd4/6e01c855-ef25-45c9-a73e-5656e8e05cd4.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 760.739044] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2672bef7-644d-4651-8488-ccb939cf9de5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.744838] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 760.744838] env[62558]: value = "task-1266564" [ 760.744838] env[62558]: _type = "Task" [ 760.744838] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.753918] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266564, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.863436] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Releasing lock "refresh_cache-3e035ee1-a63a-46a6-84ca-4980e1185e03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.863727] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.864016] env[62558]: DEBUG nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.864867] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.885617] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.139819] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.240971] env[62558]: DEBUG nova.scheduler.client.report [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.256331] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266564, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.257889] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.385767] env[62558]: DEBUG nova.network.neutron [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Updating instance_info_cache with network_info: [{"id": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "address": "fa:16:3e:87:32:7a", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd5ff8f8-6a", "ovs_interfaceid": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.387802] env[62558]: DEBUG nova.network.neutron [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.749026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.749026] env[62558]: ERROR nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Traceback (most recent call last): [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.driver.spawn(context, instance, image_meta, [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.749026] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] vm_ref = self.build_virtual_machine(instance, [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] for vif in network_info: [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return self._sync_wrapper(fn, *args, **kwargs) [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.wait() [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self[:] = self._gt.wait() [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return self._exit_event.wait() [ 761.749511] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] current.throw(*self._exc) [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] result = function(*args, **kwargs) [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] return func(*args, **kwargs) [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise e [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] nwinfo = self.network_api.allocate_for_instance( [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] created_port_ids = self._update_ports_for_instance( [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.749963] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] with excutils.save_and_reraise_exception(): [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] self.force_reraise() [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise self.value [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] updated_port = self._update_port( [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] _ensure_no_port_binding_failure(port) [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] raise exception.PortBindingFailed(port_id=port['id']) [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] nova.exception.PortBindingFailed: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. [ 761.750383] env[62558]: ERROR nova.compute.manager [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] [ 761.750715] env[62558]: DEBUG nova.compute.utils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.754348] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.865s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.758792] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Build of instance d526aa96-4007-45f8-8eee-9030bdb9f402 was re-scheduled: Binding failed for port bafe92d7-5bf8-44bf-b45b-8be12ac9aa82, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.759410] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.759790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquiring lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.760076] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Acquired lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.760364] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.769219] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266564, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516028} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.769530] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6e01c855-ef25-45c9-a73e-5656e8e05cd4/6e01c855-ef25-45c9-a73e-5656e8e05cd4.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 761.769787] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 761.770031] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-350771bb-d31a-42cd-982b-2a69f3b6b5ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.776188] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 761.776188] env[62558]: value = "task-1266565" [ 761.776188] env[62558]: _type = "Task" [ 761.776188] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.784175] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.888640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.888968] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Instance network_info: |[{"id": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "address": "fa:16:3e:87:32:7a", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd5ff8f8-6a", "ovs_interfaceid": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 761.889435] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:32:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed3ffc1d-9f86-4029-857e-6cd1d383edbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.897105] env[62558]: DEBUG oslo.service.loopingcall [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.897512] env[62558]: INFO nova.compute.manager [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] [instance: 3e035ee1-a63a-46a6-84ca-4980e1185e03] Took 1.03 seconds to deallocate network for instance. [ 761.899974] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.900412] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c859e0c-9f2b-4213-874c-9863dc1f5dc8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.921528] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.921528] env[62558]: value = "task-1266566" [ 761.921528] env[62558]: _type = "Task" [ 761.921528] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.929771] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266566, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.156245] env[62558]: DEBUG nova.compute.manager [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Received event network-changed-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.156245] env[62558]: DEBUG nova.compute.manager [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Refreshing instance network info cache due to event network-changed-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.156245] env[62558]: DEBUG oslo_concurrency.lockutils [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] Acquiring lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.191207] env[62558]: DEBUG oslo_concurrency.lockutils [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] Acquired lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.191207] env[62558]: DEBUG nova.network.neutron [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Refreshing network info cache for port bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.287563] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065114} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.288665] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.292718] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.294046] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccf4af5-08cf-4ea2-819b-7001b60759ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.321646] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 6e01c855-ef25-45c9-a73e-5656e8e05cd4/6e01c855-ef25-45c9-a73e-5656e8e05cd4.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.324491] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6317b78-a96e-4bef-aaba-119d55a8b4b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.351429] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 762.351429] env[62558]: value = "task-1266567" [ 762.351429] env[62558]: _type = "Task" [ 762.351429] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.364509] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266567, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.382143] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.435590] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266566, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.700850] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4599cbc-9ab6-460e-90dc-777169e296fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.712535] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09a0add-2402-4f89-84ba-b5fdd53b65c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.751117] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c430b4a8-7373-4b59-884c-ca44e2a49ebe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.759013] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48091c5-9c1c-4fdc-ad41-e0344bf0a7c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.774312] env[62558]: DEBUG nova.compute.provider_tree [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.865154] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.885158] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Releasing lock "refresh_cache-d526aa96-4007-45f8-8eee-9030bdb9f402" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.885446] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.885645] env[62558]: DEBUG nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.885838] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.914200] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.917375] env[62558]: DEBUG nova.network.neutron [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Updated VIF entry in instance network info cache for port bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 762.917714] env[62558]: DEBUG nova.network.neutron [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Updating instance_info_cache with network_info: [{"id": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "address": "fa:16:3e:87:32:7a", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd5ff8f8-6a", "ovs_interfaceid": "bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.936362] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266566, 'name': CreateVM_Task, 'duration_secs': 0.99059} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.936521] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.937548] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.937716] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.938114] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.939052] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56c19d28-9c3c-4cd6-9589-619471e1d492 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.944902] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 762.944902] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52586f6c-418c-b95f-0960-216101aaf271" [ 762.944902] env[62558]: _type = "Task" [ 762.944902] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.946561] env[62558]: INFO nova.scheduler.client.report [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Deleted allocations for instance 3e035ee1-a63a-46a6-84ca-4980e1185e03 [ 762.966465] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52586f6c-418c-b95f-0960-216101aaf271, 'name': SearchDatastore_Task, 'duration_secs': 0.008717} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.966788] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.967024] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.967312] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.967462] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.967638] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.967916] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54a049f7-e8ce-43dd-8644-cd7e2198b4ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.976587] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.976890] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.978183] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb51ea39-9043-4d35-997d-cc60b1a92dda {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.983585] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 762.983585] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5224d8f9-e3a8-495b-7769-7b18e7110e92" [ 762.983585] env[62558]: _type = "Task" [ 762.983585] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.992645] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5224d8f9-e3a8-495b-7769-7b18e7110e92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.278967] env[62558]: DEBUG nova.scheduler.client.report [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.363778] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266567, 'name': ReconfigVM_Task, 'duration_secs': 0.842857} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.364073] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 6e01c855-ef25-45c9-a73e-5656e8e05cd4/6e01c855-ef25-45c9-a73e-5656e8e05cd4.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.364711] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdcd7b2b-7ae6-40bf-b5af-057f84b68596 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.371147] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 763.371147] env[62558]: value = "task-1266568" [ 763.371147] env[62558]: _type = "Task" [ 763.371147] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.378562] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266568, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.422434] env[62558]: DEBUG oslo_concurrency.lockutils [req-d1db46d9-6ec2-42ce-a38a-cc7577ff9af7 req-d7dd2381-168f-4034-927d-df253afdf0fb service nova] Releasing lock "refresh_cache-b0038711-5d12-4909-b331-72acb5ed0d24" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.422971] env[62558]: DEBUG nova.network.neutron [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.462599] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46f071fa-f7a5-45cf-b340-b4e0150f0c25 tempest-ServerAddressesTestJSON-1308440485 tempest-ServerAddressesTestJSON-1308440485-project-member] Lock "3e035ee1-a63a-46a6-84ca-4980e1185e03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.896s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.496930] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5224d8f9-e3a8-495b-7769-7b18e7110e92, 'name': SearchDatastore_Task, 'duration_secs': 0.008906} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.497759] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f1acf17-824b-4e1f-be69-d11c747912bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.503062] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 763.503062] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b03c97-733a-60aa-a458-3f5808907798" [ 763.503062] env[62558]: _type = "Task" [ 763.503062] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.511030] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b03c97-733a-60aa-a458-3f5808907798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.784995] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.785714] env[62558]: ERROR nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Traceback (most recent call last): [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.driver.spawn(context, instance, image_meta, [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] vm_ref = self.build_virtual_machine(instance, [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.785714] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] for vif in network_info: [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self._sync_wrapper(fn, *args, **kwargs) [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.wait() [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self[:] = self._gt.wait() [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self._exit_event.wait() [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] result = hub.switch() [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 763.785984] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return self.greenlet.switch() [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] result = function(*args, **kwargs) [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] return func(*args, **kwargs) [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise e [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] nwinfo = self.network_api.allocate_for_instance( [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] created_port_ids = self._update_ports_for_instance( [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] with excutils.save_and_reraise_exception(): [ 763.786308] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] self.force_reraise() [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise self.value [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] updated_port = self._update_port( [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] _ensure_no_port_binding_failure(port) [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] raise exception.PortBindingFailed(port_id=port['id']) [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] nova.exception.PortBindingFailed: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. [ 763.786615] env[62558]: ERROR nova.compute.manager [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] [ 763.786894] env[62558]: DEBUG nova.compute.utils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.787801] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.488s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.790632] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Build of instance d92dca09-0a42-4103-ac92-25b6b0fa3c32 was re-scheduled: Binding failed for port d634a7b0-51b7-466d-8e64-2f75affbe7f9, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.791209] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 763.791452] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.791598] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.791755] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.882302] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266568, 'name': Rename_Task, 'duration_secs': 0.197096} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.882603] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.882871] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04d7a77c-f378-4d47-9df2-ed3254250b2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.890484] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 763.890484] env[62558]: value = "task-1266569" [ 763.890484] env[62558]: _type = "Task" [ 763.890484] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.901113] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.925920] env[62558]: INFO nova.compute.manager [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] [instance: d526aa96-4007-45f8-8eee-9030bdb9f402] Took 1.04 seconds to deallocate network for instance. [ 763.965162] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.015515] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b03c97-733a-60aa-a458-3f5808907798, 'name': SearchDatastore_Task, 'duration_secs': 0.010175} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.015792] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.016276] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] b0038711-5d12-4909-b331-72acb5ed0d24/b0038711-5d12-4909-b331-72acb5ed0d24.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.016394] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e21b965-2935-47a8-bbf1-aadaa8f1248c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.024397] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 764.024397] env[62558]: value = "task-1266570" [ 764.024397] env[62558]: _type = "Task" [ 764.024397] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.036083] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266570, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.319573] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.401869] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266569, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.495747] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.543405] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266570, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.590201] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.603165] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "80221843-4e15-4f20-aeb4-4e6081371b95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.603412] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.714964] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209756fa-ce18-47f3-b75d-8fce361e1de4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.722513] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9654d28-6ab1-4260-9358-061d43cbdbd9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.754184] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430e665d-396e-448b-902d-bd9fad91b76d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.761767] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83eb6c9e-70ad-4fce-9559-7f6f6af789f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.774814] env[62558]: DEBUG nova.compute.provider_tree [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.901302] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266569, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.960326] env[62558]: INFO nova.scheduler.client.report [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Deleted allocations for instance d526aa96-4007-45f8-8eee-9030bdb9f402 [ 765.038623] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266570, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540353} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.039153] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] b0038711-5d12-4909-b331-72acb5ed0d24/b0038711-5d12-4909-b331-72acb5ed0d24.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.042022] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.042022] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7c7a1c9-a7a8-4589-88bf-7847319151cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.049018] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 765.049018] env[62558]: value = "task-1266571" [ 765.049018] env[62558]: _type = "Task" [ 765.049018] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.060030] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.095422] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Releasing lock "refresh_cache-d92dca09-0a42-4103-ac92-25b6b0fa3c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.095422] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.095593] env[62558]: DEBUG nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.095764] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.117286] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.280941] env[62558]: DEBUG nova.scheduler.client.report [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.404024] env[62558]: DEBUG oslo_vmware.api [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266569, 'name': PowerOnVM_Task, 'duration_secs': 1.244792} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.404747] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 765.405122] env[62558]: INFO nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Took 10.05 seconds to spawn the instance on the hypervisor. [ 765.405429] env[62558]: DEBUG nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 765.406310] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f887d27c-84a8-4976-98b7-7fd3a828bbd7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.470905] env[62558]: DEBUG oslo_concurrency.lockutils [None req-701f1fd4-d498-438b-94cf-1d114278423e tempest-ServerRescueTestJSON-1756443674 tempest-ServerRescueTestJSON-1756443674-project-member] Lock "d526aa96-4007-45f8-8eee-9030bdb9f402" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.155s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.560176] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079142} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.561326] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.562090] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b62ee7c-eb18-4b8d-a099-bc9e869a10c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.586384] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] b0038711-5d12-4909-b331-72acb5ed0d24/b0038711-5d12-4909-b331-72acb5ed0d24.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.586740] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dca56de8-5df6-404f-9510-4988c78766ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.611324] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 765.611324] env[62558]: value = "task-1266572" [ 765.611324] env[62558]: _type = "Task" [ 765.611324] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.619861] env[62558]: DEBUG nova.network.neutron [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.621051] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266572, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.783524] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.784195] env[62558]: ERROR nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Traceback (most recent call last): [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.driver.spawn(context, instance, image_meta, [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] vm_ref = self.build_virtual_machine(instance, [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.784195] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] for vif in network_info: [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return self._sync_wrapper(fn, *args, **kwargs) [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.wait() [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self[:] = self._gt.wait() [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return self._exit_event.wait() [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] current.throw(*self._exc) [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.784459] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] result = function(*args, **kwargs) [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] return func(*args, **kwargs) [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise e [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] nwinfo = self.network_api.allocate_for_instance( [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] created_port_ids = self._update_ports_for_instance( [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] with excutils.save_and_reraise_exception(): [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] self.force_reraise() [ 765.784711] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise self.value [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] updated_port = self._update_port( [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] _ensure_no_port_binding_failure(port) [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] raise exception.PortBindingFailed(port_id=port['id']) [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] nova.exception.PortBindingFailed: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. [ 765.784966] env[62558]: ERROR nova.compute.manager [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] [ 765.784966] env[62558]: DEBUG nova.compute.utils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.786699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.150s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.788256] env[62558]: INFO nova.compute.claims [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.790960] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Build of instance f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0 was re-scheduled: Binding failed for port a9fb6ae9-17b1-45d9-b2ad-0e2ac6e8bf7b, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.791485] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.791782] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquiring lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.792018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Acquired lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.796027] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.924928] env[62558]: INFO nova.compute.manager [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Took 31.69 seconds to build instance. [ 765.973406] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.126085] env[62558]: INFO nova.compute.manager [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: d92dca09-0a42-4103-ac92-25b6b0fa3c32] Took 1.03 seconds to deallocate network for instance. [ 766.127533] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266572, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.322780] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.426967] env[62558]: DEBUG oslo_concurrency.lockutils [None req-af29be1b-de1c-4e41-990e-a998b89f369e tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.807s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.451869] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.502632] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.625916] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266572, 'name': ReconfigVM_Task, 'duration_secs': 0.760053} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.626239] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Reconfigured VM instance instance-00000035 to attach disk [datastore2] b0038711-5d12-4909-b331-72acb5ed0d24/b0038711-5d12-4909-b331-72acb5ed0d24.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.630969] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4a529a9-33fe-46bb-a84f-f4964a69c9d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.641071] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 766.641071] env[62558]: value = "task-1266573" [ 766.641071] env[62558]: _type = "Task" [ 766.641071] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.654218] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266573, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.934804] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.955772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Releasing lock "refresh_cache-f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.955772] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.955772] env[62558]: DEBUG nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.955772] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.978717] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.149391] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266573, 'name': Rename_Task, 'duration_secs': 0.159067} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.150234] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.150234] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f63a0c61-5e5d-4456-b245-7555eeab76b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.157046] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 767.157046] env[62558]: value = "task-1266574" [ 767.157046] env[62558]: _type = "Task" [ 767.157046] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.162180] env[62558]: INFO nova.scheduler.client.report [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Deleted allocations for instance d92dca09-0a42-4103-ac92-25b6b0fa3c32 [ 767.177823] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266574, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.321724] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbe6395-3292-4e35-9be8-2444f779f082 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.329389] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f06f26-3e78-439a-984d-c8f53d989d5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.364343] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb649ea2-524a-460f-80ef-8753bbabdc8d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.372319] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17eb5db-cdaa-4483-92d9-6c38b8ac3885 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.387893] env[62558]: DEBUG nova.compute.provider_tree [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.450703] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.481045] env[62558]: DEBUG nova.network.neutron [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.666996] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266574, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.678665] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1ca9fae2-19d5-465a-a0ca-e29f6f094362 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "d92dca09-0a42-4103-ac92-25b6b0fa3c32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.194s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.892260] env[62558]: DEBUG nova.scheduler.client.report [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.983825] env[62558]: INFO nova.compute.manager [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] [instance: f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0] Took 1.03 seconds to deallocate network for instance. [ 768.169122] env[62558]: DEBUG oslo_vmware.api [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266574, 'name': PowerOnVM_Task, 'duration_secs': 0.917841} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.169471] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 768.169767] env[62558]: INFO nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Took 8.42 seconds to spawn the instance on the hypervisor. [ 768.169982] env[62558]: DEBUG nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 768.170807] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374f9761-7b0e-4b9f-8ce4-c451f132f80c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.181324] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.397666] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.398073] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.404254] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.999s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.710972] env[62558]: INFO nova.compute.manager [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Took 29.92 seconds to build instance. [ 768.721493] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.906803] env[62558]: DEBUG nova.compute.utils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.911929] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.912132] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.984721] env[62558]: DEBUG nova.policy [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd3a894e5784481eaff0fdd0fa414cee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b9d93976e6045c288ad6b152304ef72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.020466] env[62558]: INFO nova.scheduler.client.report [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Deleted allocations for instance f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0 [ 769.213933] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46fc42bc-527d-444d-aa8a-8f60ff627bcc tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.173s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.387462] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Successfully created port: d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.410294] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.442205] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d65fbfe3-b1e9-48d9-88ba-209fdc264283 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.442402] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.442527] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6e01c855-ef25-45c9-a73e-5656e8e05cd4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.442650] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance b0038711-5d12-4909-b331-72acb5ed0d24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.444480] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ac801d7-af70-46e0-88b3-02caee13497d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.530531] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f806161-e050-4746-aa68-d435b52a0fb3 tempest-SecurityGroupsTestJSON-785116825 tempest-SecurityGroupsTestJSON-785116825-project-member] Lock "f4dfa1c6-6c8b-4098-b3a2-9b229fc57bb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.938s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.717392] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.946240] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.033731] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.245184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.425964] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.448718] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 8acd06fc-b040-468c-980f-6e17b2343c4a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.455424] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.455693] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.455851] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.458141] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.458385] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.458555] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.458772] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.458940] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.459131] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.459298] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.459467] env[62558]: DEBUG nova.virt.hardware [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.460413] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94840927-ccf2-4b56-9b9b-b9201fa6e742 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.469347] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa9531f-f305-4817-a4ca-06afb04b54aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.561083] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.897858] env[62558]: DEBUG nova.compute.manager [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Received event network-vif-plugged-d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.898132] env[62558]: DEBUG oslo_concurrency.lockutils [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] Acquiring lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.898325] env[62558]: DEBUG oslo_concurrency.lockutils [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] Lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.898490] env[62558]: DEBUG oslo_concurrency.lockutils [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] Lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.898654] env[62558]: DEBUG nova.compute.manager [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] No waiting events found dispatching network-vif-plugged-d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 770.898813] env[62558]: WARNING nova.compute.manager [req-ba1f8296-733c-45fb-9318-4d83bf6ced6d req-dcb043d8-843b-4769-8a72-2471e908cd89 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Received unexpected event network-vif-plugged-d9a83e75-1cf0-4747-90ee-2fec31fff5ad for instance with vm_state building and task_state spawning. [ 770.915985] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Successfully updated port: d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 770.964462] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 3c8c2d25-981e-49de-b5c0-e0d12dd91378 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.995927] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.996681] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.421707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.421707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.421707] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.468849] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a8fef137-a651-4724-8ad1-a3e6ddcb5b03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.903519] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.903937] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.904052] env[62558]: DEBUG nova.compute.manager [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 771.904984] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac521b15-9a67-4fda-851a-347110a73d5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.912923] env[62558]: DEBUG nova.compute.manager [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 771.913615] env[62558]: DEBUG nova.objects.instance [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'flavor' on Instance uuid 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 771.963102] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.972425] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6d89107f-0727-4b8c-863d-d1e635000bff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.127946] env[62558]: DEBUG nova.network.neutron [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.422642] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.422642] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6f094a3-0401-4575-b66f-53759ebb01d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.430021] env[62558]: DEBUG oslo_vmware.api [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 772.430021] env[62558]: value = "task-1266575" [ 772.430021] env[62558]: _type = "Task" [ 772.430021] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.439172] env[62558]: DEBUG oslo_vmware.api [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.479782] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a1d242e6-1561-4bd4-8e39-281ab6346661 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.634037] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.634037] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Instance network_info: |[{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 772.634300] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:79:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9a83e75-1cf0-4747-90ee-2fec31fff5ad', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.640265] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating folder: Project (0b9d93976e6045c288ad6b152304ef72). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.640878] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bcf1afd-66d4-4a00-9524-970fa9b71f01 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.653127] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created folder: Project (0b9d93976e6045c288ad6b152304ef72) in parent group-v272451. [ 772.653127] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating folder: Instances. Parent ref: group-v272477. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.653127] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abcc7425-9a57-482e-b7dd-be974ead71d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.660057] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created folder: Instances in parent group-v272477. [ 772.660516] env[62558]: DEBUG oslo.service.loopingcall [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.661425] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.661766] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6ac8765-5bd9-435e-9404-f84dd5a15aed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.681680] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.681680] env[62558]: value = "task-1266578" [ 772.681680] env[62558]: _type = "Task" [ 772.681680] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.689976] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266578, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.935373] env[62558]: DEBUG nova.compute.manager [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Received event network-changed-d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.935632] env[62558]: DEBUG nova.compute.manager [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Refreshing instance network info cache due to event network-changed-d9a83e75-1cf0-4747-90ee-2fec31fff5ad. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.935821] env[62558]: DEBUG oslo_concurrency.lockutils [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] Acquiring lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.935960] env[62558]: DEBUG oslo_concurrency.lockutils [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] Acquired lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.936396] env[62558]: DEBUG nova.network.neutron [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Refreshing network info cache for port d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.944196] env[62558]: DEBUG oslo_vmware.api [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266575, 'name': PowerOffVM_Task, 'duration_secs': 0.220388} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.944664] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 772.948037] env[62558]: DEBUG nova.compute.manager [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.948037] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c55cb0-8ec0-4539-90d3-d8ccd73eb0d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.983756] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.194981] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266578, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.464669] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b57c4a28-8693-4aa1-9a3f-704878cecf68 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.561s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.488017] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ebd84b3f-0c5d-40e3-aa70-f8b3054b7109 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.692492] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266578, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.913225] env[62558]: DEBUG nova.network.neutron [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updated VIF entry in instance network info cache for port d9a83e75-1cf0-4747-90ee-2fec31fff5ad. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 773.913470] env[62558]: DEBUG nova.network.neutron [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.991445] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.201160] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266578, 'name': CreateVM_Task, 'duration_secs': 1.346007} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.201160] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 774.201901] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.201901] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.203013] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 774.203013] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d120275-1132-4da5-b50f-c72766d9f9d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.211019] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 774.211019] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52efa725-ced1-4e48-876e-6d684be9e412" [ 774.211019] env[62558]: _type = "Task" [ 774.211019] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.217472] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52efa725-ced1-4e48-876e-6d684be9e412, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.416637] env[62558]: DEBUG oslo_concurrency.lockutils [req-1fe4cb0a-1482-4955-a773-42a1cb7e0782 req-3b8c0bb9-48f7-460d-8b8e-f3718efcd7f0 service nova] Releasing lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.494894] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.695596] env[62558]: DEBUG nova.objects.instance [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'flavor' on Instance uuid 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 774.720286] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52efa725-ced1-4e48-876e-6d684be9e412, 'name': SearchDatastore_Task, 'duration_secs': 0.010087} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.720637] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.720887] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.721126] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.721269] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.721441] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.721700] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06fe5d21-3ac8-4bf2-9322-fef483f7aeb3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.729644] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.729829] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 774.730846] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28be18d3-1ff8-4c75-9d4d-0222e21e3190 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.736223] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 774.736223] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bf1be0-f630-eb9e-5fed-dda1707e4c07" [ 774.736223] env[62558]: _type = "Task" [ 774.736223] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.743799] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bf1be0-f630-eb9e-5fed-dda1707e4c07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.999389] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4a3f7642-1b9c-40d8-973b-5153b559bda0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.201732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.201907] env[62558]: DEBUG oslo_concurrency.lockutils [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquired lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.202183] env[62558]: DEBUG nova.network.neutron [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.202412] env[62558]: DEBUG nova.objects.instance [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'info_cache' on Instance uuid 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 775.247940] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bf1be0-f630-eb9e-5fed-dda1707e4c07, 'name': SearchDatastore_Task, 'duration_secs': 0.00783} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.249037] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7912e248-6c02-43da-8204-a9fabf2fa938 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.253954] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 775.253954] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d2090-64b5-a727-c270-d0d7083feb40" [ 775.253954] env[62558]: _type = "Task" [ 775.253954] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.261225] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d2090-64b5-a727-c270-d0d7083feb40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.502630] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance e846e43d-68ec-4de3-ba62-e538643b6e4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.705977] env[62558]: DEBUG nova.objects.base [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Object Instance<867c0fb3-d4f6-4945-a5cc-6d93efcadb79> lazy-loaded attributes: flavor,info_cache {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 775.764240] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d2090-64b5-a727-c270-d0d7083feb40, 'name': SearchDatastore_Task, 'duration_secs': 0.009917} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.764437] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.764721] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/2ac801d7-af70-46e0-88b3-02caee13497d.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 775.764990] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b668ffa2-e5ae-4425-b9a8-08f586b0f482 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.771163] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 775.771163] env[62558]: value = "task-1266579" [ 775.771163] env[62558]: _type = "Task" [ 775.771163] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.780422] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.007356] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4526df2a-f801-4de8-8218-497a7d22034f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 776.282242] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457414} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.282484] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/2ac801d7-af70-46e0-88b3-02caee13497d.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 776.282695] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.282937] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d0d8a62-af67-4f5d-8ba1-1a919eca91b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.289500] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 776.289500] env[62558]: value = "task-1266580" [ 776.289500] env[62558]: _type = "Task" [ 776.289500] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.299500] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.424666] env[62558]: DEBUG nova.network.neutron [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updating instance_info_cache with network_info: [{"id": "99b91852-f90d-4a37-9558-042993918fd5", "address": "fa:16:3e:7b:40:aa", "network": {"id": "9728862e-95d3-4dbf-875e-4035afa5d290", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1695917110-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e51dcdbfe744ddf909800d2cdc1b98c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed3ffc1d-9f86-4029-857e-6cd1d383edbb", "external-id": "nsx-vlan-transportzone-759", "segmentation_id": 759, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99b91852-f9", "ovs_interfaceid": "99b91852-f90d-4a37-9558-042993918fd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.510127] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 49a58b46-207f-4515-b313-afcdb2d1ced3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 776.805065] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06511} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.805489] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.806650] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ae23a2-872f-4183-b37b-d5923f3befd8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.829188] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/2ac801d7-af70-46e0-88b3-02caee13497d.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.829485] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75f8d7be-664f-4281-8be0-954e3a3a96b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.848127] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 776.848127] env[62558]: value = "task-1266581" [ 776.848127] env[62558]: _type = "Task" [ 776.848127] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.855263] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.927239] env[62558]: DEBUG oslo_concurrency.lockutils [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Releasing lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.013027] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 96361b52-b939-4409-aaf7-ba5b3bf054e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.358245] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266581, 'name': ReconfigVM_Task, 'duration_secs': 0.296625} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.358522] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/2ac801d7-af70-46e0-88b3-02caee13497d.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.359161] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63cc738d-9199-4da1-8a3d-164aaf98d41b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.365443] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 777.365443] env[62558]: value = "task-1266582" [ 777.365443] env[62558]: _type = "Task" [ 777.365443] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.373041] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266582, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.432373] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.432699] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f12def98-4de4-4a61-b369-c21151c3f1cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.439640] env[62558]: DEBUG oslo_vmware.api [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 777.439640] env[62558]: value = "task-1266583" [ 777.439640] env[62558]: _type = "Task" [ 777.439640] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.447138] env[62558]: DEBUG oslo_vmware.api [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.516211] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 66ff3b25-d49c-4129-8b51-5338e75b09b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.876943] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266582, 'name': Rename_Task, 'duration_secs': 0.128344} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.877245] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.877489] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99fde435-13fc-4d93-bc48-66ec178599e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.883153] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 777.883153] env[62558]: value = "task-1266584" [ 777.883153] env[62558]: _type = "Task" [ 777.883153] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.598952] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6243631f-b240-4d7c-8910-0bf3d2dedb77 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 778.607710] env[62558]: DEBUG oslo_vmware.api [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266584, 'name': PowerOnVM_Task, 'duration_secs': 0.434663} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.608322] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.608549] env[62558]: INFO nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Took 8.18 seconds to spawn the instance on the hypervisor. [ 778.608750] env[62558]: DEBUG nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.609514] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8251a0c6-1643-4fe9-ac5e-f0b76357cc39 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.615378] env[62558]: DEBUG oslo_vmware.api [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266583, 'name': PowerOnVM_Task, 'duration_secs': 0.423943} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.615922] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.616140] env[62558]: DEBUG nova.compute.manager [None req-dbfe2cc8-aa23-4019-8948-df209a65a640 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.617106] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cdd59b-9fa2-4b0f-ab75-ba31f751cfea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.102410] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 779.135822] env[62558]: INFO nova.compute.manager [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Took 28.51 seconds to build instance. [ 779.606647] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 779.637695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d770e462-2cbc-4e8c-b786-d1fe3258cbd8 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.473s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.110430] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 80221843-4e15-4f20-aeb4-4e6081371b95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.110625] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 780.110776] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 780.140670] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.180895] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 780.195741] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 780.195928] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.208148] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 780.225853] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 780.564770] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce596ae-e1ff-4cf4-a802-5298fcf8370d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.572497] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c277fd-445c-435e-9222-fe889de66d51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.602357] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44726738-2761-4f25-af59-ba49bba20208 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.609197] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dabb65-3bcc-4ba9-a7fc-fd9e9607b769 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.622462] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.658710] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.978723] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "b0038711-5d12-4909-b331-72acb5ed0d24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.979050] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.979311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.979535] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.979736] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.981933] env[62558]: INFO nova.compute.manager [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Terminating instance [ 780.983954] env[62558]: DEBUG nova.compute.manager [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.984169] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.984988] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc784fad-3e7d-480e-85d7-ffc5f21ec901 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.992627] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.992864] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ff1bed6-936a-4c6b-8af8-1845157cb8e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.000232] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 781.000232] env[62558]: value = "task-1266585" [ 781.000232] env[62558]: _type = "Task" [ 781.000232] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.009686] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.125215] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.510063] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266585, 'name': PowerOffVM_Task, 'duration_secs': 0.177898} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.510341] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.510508] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.510753] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dabf888a-7082-48fd-8b96-83e6c4aebae2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.579040] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.579174] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.579302] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleting the datastore file [datastore2] b0038711-5d12-4909-b331-72acb5ed0d24 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.579567] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79765bcd-1624-4216-ae66-2a713d2dc8f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.591726] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 781.591726] env[62558]: value = "task-1266587" [ 781.591726] env[62558]: _type = "Task" [ 781.591726] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.598520] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.631370] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 781.631370] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.228s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.631370] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.888s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.634786] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 781.634786] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 782.101285] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.142864] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] There are 4 instances to clean {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 782.142864] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f4f7e2c9-229c-4547-b674-5a5219a28151] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 782.516674] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f884c1-da31-428d-bc75-afb13dd0b8e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.523671] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf2086e-8b6d-4113-be3c-5de8a23d76bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.552793] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4f0ef5-a45c-4fd2-8973-df889a929680 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.559135] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0febc3b2-f7ad-494e-97ed-ae460bea8843 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.571485] env[62558]: DEBUG nova.compute.provider_tree [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.601682] env[62558]: DEBUG oslo_vmware.api [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.581648} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.601917] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.602130] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.602307] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.602509] env[62558]: INFO nova.compute.manager [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Took 1.62 seconds to destroy the instance on the hypervisor. [ 782.602746] env[62558]: DEBUG oslo.service.loopingcall [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.602929] env[62558]: DEBUG nova.compute.manager [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.603033] env[62558]: DEBUG nova.network.neutron [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.648111] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9316facd-0c95-4720-b0d3-c0a39df81619] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 782.887401] env[62558]: DEBUG nova.compute.manager [req-5145cbd4-6e0f-444d-9110-92af7b592752 req-3690df63-373a-4e07-a773-e38d5b865a80 service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Received event network-vif-deleted-bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.887401] env[62558]: INFO nova.compute.manager [req-5145cbd4-6e0f-444d-9110-92af7b592752 req-3690df63-373a-4e07-a773-e38d5b865a80 service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Neutron deleted interface bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498; detaching it from the instance and deleting it from the info cache [ 782.887569] env[62558]: DEBUG nova.network.neutron [req-5145cbd4-6e0f-444d-9110-92af7b592752 req-3690df63-373a-4e07-a773-e38d5b865a80 service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.080037] env[62558]: DEBUG nova.scheduler.client.report [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.151479] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d02e6383-42d9-40db-89a3-ba8b5595b95b] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 783.354411] env[62558]: DEBUG nova.network.neutron [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.390776] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-440576c2-f0e6-44ad-a689-81aa0a8fae10 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.400608] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ec48ca-57ef-4b39-b17d-99528150cd1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.424234] env[62558]: DEBUG nova.compute.manager [req-5145cbd4-6e0f-444d-9110-92af7b592752 req-3690df63-373a-4e07-a773-e38d5b865a80 service nova] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Detach interface failed, port_id=bd5ff8f8-6a94-4a9c-8cab-7643aa2fa498, reason: Instance b0038711-5d12-4909-b331-72acb5ed0d24 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 783.582467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.583123] env[62558]: ERROR nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Traceback (most recent call last): [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.driver.spawn(context, instance, image_meta, [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] vm_ref = self.build_virtual_machine(instance, [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.583123] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] for vif in network_info: [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return self._sync_wrapper(fn, *args, **kwargs) [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.wait() [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self[:] = self._gt.wait() [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return self._exit_event.wait() [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] current.throw(*self._exc) [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.583439] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] result = function(*args, **kwargs) [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] return func(*args, **kwargs) [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise e [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] nwinfo = self.network_api.allocate_for_instance( [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] created_port_ids = self._update_ports_for_instance( [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] with excutils.save_and_reraise_exception(): [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] self.force_reraise() [ 783.583795] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise self.value [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] updated_port = self._update_port( [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] _ensure_no_port_binding_failure(port) [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] raise exception.PortBindingFailed(port_id=port['id']) [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] nova.exception.PortBindingFailed: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. [ 783.584208] env[62558]: ERROR nova.compute.manager [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] [ 783.584208] env[62558]: DEBUG nova.compute.utils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. {{(pid=62558) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.585079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.582s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.586494] env[62558]: INFO nova.compute.claims [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.589097] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Build of instance d65fbfe3-b1e9-48d9-88ba-209fdc264283 was re-scheduled: Binding failed for port 1f8840f2-1974-4bc1-95e6-0e5c6d1d30cd, please check neutron logs for more information. {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.589509] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Unplugging VIFs for instance {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.589730] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquiring lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.589872] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Acquired lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.590074] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.657956] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 06df57a7-2c53-4f79-bec4-e46cfe1dca63] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 783.857418] env[62558]: INFO nova.compute.manager [-] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Took 1.25 seconds to deallocate network for instance. [ 784.113471] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.162435] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 784.162613] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances with incomplete migration {{(pid=62558) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 784.202284] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.364670] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.664691] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 784.705416] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Releasing lock "refresh_cache-d65fbfe3-b1e9-48d9-88ba-209fdc264283" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.705782] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62558) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.706090] env[62558]: DEBUG nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.706398] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.723854] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.885761] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32978ab-d878-4b17-867b-fb99c1663bfe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.893333] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b965695-b3c8-41a5-a872-37145565eb47 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.922452] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab5f188-f31e-4b48-b552-932c4ff2c241 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.929781] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c783ae36-3054-4215-a38f-6aecb1914ea2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.942877] env[62558]: DEBUG nova.compute.provider_tree [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.226465] env[62558]: DEBUG nova.network.neutron [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.448400] env[62558]: DEBUG nova.scheduler.client.report [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.731165] env[62558]: INFO nova.compute.manager [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] [instance: d65fbfe3-b1e9-48d9-88ba-209fdc264283] Took 1.02 seconds to deallocate network for instance. [ 785.953859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.954982] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.958146] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.818s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.959804] env[62558]: INFO nova.compute.claims [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.466612] env[62558]: DEBUG nova.compute.utils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.468164] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.468522] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.512129] env[62558]: DEBUG nova.policy [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd3a894e5784481eaff0fdd0fa414cee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b9d93976e6045c288ad6b152304ef72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.764952] env[62558]: INFO nova.scheduler.client.report [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Deleted allocations for instance d65fbfe3-b1e9-48d9-88ba-209fdc264283 [ 786.785298] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Successfully created port: 15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.971632] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.251146] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc1e1d5-4760-47e3-b7e2-f7b6fef950c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.258576] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d55413e-d68c-4ae6-ad8b-4445d7dcdc3e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.288785] env[62558]: DEBUG oslo_concurrency.lockutils [None req-29dec96f-770b-4bad-8ee5-2e76a29903df tempest-VolumesAdminNegativeTest-1780069695 tempest-VolumesAdminNegativeTest-1780069695-project-member] Lock "d65fbfe3-b1e9-48d9-88ba-209fdc264283" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 165.512s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.291975] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66853bc2-f0aa-476f-a5a4-9499ac88e7fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.300192] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96767d4-fb81-4a6e-9100-9c0f304e9ed0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.314064] env[62558]: DEBUG nova.compute.provider_tree [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.791979] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.816900] env[62558]: DEBUG nova.scheduler.client.report [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.983066] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.009107] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.009384] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.009553] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.009734] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.009901] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.010064] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.010385] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.010550] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.010739] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.010904] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.011097] env[62558]: DEBUG nova.virt.hardware [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.011985] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a311e515-7b1d-4f86-95c7-cb5a4d051b95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.020063] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce17d347-8c88-48ea-a0e6-24321349b521 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.191140] env[62558]: DEBUG nova.compute.manager [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Received event network-vif-plugged-15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.191426] env[62558]: DEBUG oslo_concurrency.lockutils [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] Acquiring lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.191652] env[62558]: DEBUG oslo_concurrency.lockutils [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.191831] env[62558]: DEBUG oslo_concurrency.lockutils [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.191967] env[62558]: DEBUG nova.compute.manager [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] No waiting events found dispatching network-vif-plugged-15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.192310] env[62558]: WARNING nova.compute.manager [req-0e0fcf97-977d-455c-94c0-b1c6abe18258 req-48fdf8a7-154f-43e6-9e46-8557775b5580 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Received unexpected event network-vif-plugged-15338668-abe5-414d-91c8-3396fdaf1b3e for instance with vm_state building and task_state spawning. [ 788.286493] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Successfully updated port: 15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.319487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.324309] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.324309] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.325617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.830s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.327331] env[62558]: INFO nova.compute.claims [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.790184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.790391] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.790555] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.832402] env[62558]: DEBUG nova.compute.utils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.835876] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.836050] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.876982] env[62558]: DEBUG nova.policy [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18ef6cfb3b824623b20cbeb886439695', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd61c731af2b5475f8d236b7d31b6073d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.161982] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Successfully created port: 23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.330912] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.340884] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.532382] env[62558]: DEBUG nova.network.neutron [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [{"id": "15338668-abe5-414d-91c8-3396fdaf1b3e", "address": "fa:16:3e:f8:9c:60", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15338668-ab", "ovs_interfaceid": "15338668-abe5-414d-91c8-3396fdaf1b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.730680] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342b1e60-e574-4786-8a7a-99fc72dae1ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.738985] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c057cf-3184-4cca-b865-ca55652e96bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.772255] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da8bbae-5e70-4860-a900-f6d75f2124d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.781130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6d61eb-e464-4efa-9829-bd179aa42c2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.793796] env[62558]: DEBUG nova.compute.provider_tree [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.036074] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.036435] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance network_info: |[{"id": "15338668-abe5-414d-91c8-3396fdaf1b3e", "address": "fa:16:3e:f8:9c:60", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15338668-ab", "ovs_interfaceid": "15338668-abe5-414d-91c8-3396fdaf1b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.036890] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:9c:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15338668-abe5-414d-91c8-3396fdaf1b3e', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.045438] env[62558]: DEBUG oslo.service.loopingcall [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.045669] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.045887] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c59f8d3c-9641-482e-8069-c8773cfe81e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.064807] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.064807] env[62558]: value = "task-1266588" [ 790.064807] env[62558]: _type = "Task" [ 790.064807] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.073628] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266588, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.217626] env[62558]: DEBUG nova.compute.manager [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Received event network-changed-15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.217945] env[62558]: DEBUG nova.compute.manager [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Refreshing instance network info cache due to event network-changed-15338668-abe5-414d-91c8-3396fdaf1b3e. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.218252] env[62558]: DEBUG oslo_concurrency.lockutils [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] Acquiring lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.218473] env[62558]: DEBUG oslo_concurrency.lockutils [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] Acquired lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.218688] env[62558]: DEBUG nova.network.neutron [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Refreshing network info cache for port 15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 790.299758] env[62558]: DEBUG nova.scheduler.client.report [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.351057] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.379773] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.380034] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.380207] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.380418] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.380570] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.380717] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.381036] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.381228] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.381402] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.381562] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.381730] env[62558]: DEBUG nova.virt.hardware [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.382909] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d670b52-6276-4e8d-8e3d-8f1102bef44d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.390369] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6364c727-d705-4bd4-8283-99dfcf82f9f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.577191] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266588, 'name': CreateVM_Task, 'duration_secs': 0.340122} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.577368] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.578054] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.578222] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.578563] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.578823] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2327aea9-d6f4-4a4b-8651-260d19a4fa4e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.583984] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 790.583984] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d5ae3b-0111-927c-adbf-302e74d27d14" [ 790.583984] env[62558]: _type = "Task" [ 790.583984] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.595339] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d5ae3b-0111-927c-adbf-302e74d27d14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.811502] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.811502] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.814584] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.312s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.816726] env[62558]: INFO nova.compute.claims [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.904958] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Successfully updated port: 23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.095076] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d5ae3b-0111-927c-adbf-302e74d27d14, 'name': SearchDatastore_Task, 'duration_secs': 0.011536} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.095347] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.095486] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.095716] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.095860] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.096087] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.096307] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de53d3fd-2c1b-4570-8b3d-0697966083b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.107437] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.107691] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.114213] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdb729de-f966-439b-beb3-e7a0b56b1870 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.121443] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 791.121443] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52816ba9-7c8b-00c5-e660-966bb1d2f2f9" [ 791.121443] env[62558]: _type = "Task" [ 791.121443] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.128150] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52816ba9-7c8b-00c5-e660-966bb1d2f2f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.174191] env[62558]: DEBUG nova.network.neutron [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updated VIF entry in instance network info cache for port 15338668-abe5-414d-91c8-3396fdaf1b3e. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 791.174582] env[62558]: DEBUG nova.network.neutron [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [{"id": "15338668-abe5-414d-91c8-3396fdaf1b3e", "address": "fa:16:3e:f8:9c:60", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15338668-ab", "ovs_interfaceid": "15338668-abe5-414d-91c8-3396fdaf1b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.324496] env[62558]: DEBUG nova.compute.utils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.328712] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.328862] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.386956] env[62558]: DEBUG nova.policy [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9769ac5368ed412c90eba91ec6c66401', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3625580997514184ab36f8258cf2358b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.408859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.409034] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquired lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.409219] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.635341] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52816ba9-7c8b-00c5-e660-966bb1d2f2f9, 'name': SearchDatastore_Task, 'duration_secs': 0.008311} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.635341] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ea747f0-a864-4635-b057-1da382f4ca39 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.641149] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 791.641149] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f108dc-4a8e-9d79-9bba-bc60009fb3a5" [ 791.641149] env[62558]: _type = "Task" [ 791.641149] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.650287] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f108dc-4a8e-9d79-9bba-bc60009fb3a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.651074] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Successfully created port: c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.677980] env[62558]: DEBUG oslo_concurrency.lockutils [req-a88e7ffa-7c48-4544-a6d3-f52502960e59 req-bd072095-44dd-4930-b3e0-0d0e571b2b27 service nova] Releasing lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.829685] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.956405] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.146875] env[62558]: DEBUG nova.network.neutron [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Updating instance_info_cache with network_info: [{"id": "23817056-3d15-4a89-8f18-84e4f7549efe", "address": "fa:16:3e:83:80:3d", "network": {"id": "c2b530e3-a253-4340-ba85-a76cf297211f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1520162603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61c731af2b5475f8d236b7d31b6073d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23817056-3d", "ovs_interfaceid": "23817056-3d15-4a89-8f18-84e4f7549efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.157018] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f108dc-4a8e-9d79-9bba-bc60009fb3a5, 'name': SearchDatastore_Task, 'duration_secs': 0.00959} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.157018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.157153] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.157635] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-001b81f1-818c-4eef-ad03-75a38cfad950 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.165576] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 792.165576] env[62558]: value = "task-1266589" [ 792.165576] env[62558]: _type = "Task" [ 792.165576] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.176513] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266589, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.227697] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4e2537-ceca-4864-8f1e-22e019af7a18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.235636] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ad9757-5db4-4795-93a2-3efae1b46fe0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.269467] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1b1dcc-4118-425c-a6b6-9541b7019775 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.277186] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bcd244-7a6e-4771-beb8-fb7dc3a97ac2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.291285] env[62558]: DEBUG nova.compute.provider_tree [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.326555] env[62558]: DEBUG nova.compute.manager [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Received event network-vif-plugged-23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.326889] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Acquiring lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.327248] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.327421] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.327694] env[62558]: DEBUG nova.compute.manager [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] No waiting events found dispatching network-vif-plugged-23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 792.327908] env[62558]: WARNING nova.compute.manager [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Received unexpected event network-vif-plugged-23817056-3d15-4a89-8f18-84e4f7549efe for instance with vm_state building and task_state spawning. [ 792.328089] env[62558]: DEBUG nova.compute.manager [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Received event network-changed-23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.328261] env[62558]: DEBUG nova.compute.manager [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Refreshing instance network info cache due to event network-changed-23817056-3d15-4a89-8f18-84e4f7549efe. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.328435] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Acquiring lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.650082] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Releasing lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.650523] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Instance network_info: |[{"id": "23817056-3d15-4a89-8f18-84e4f7549efe", "address": "fa:16:3e:83:80:3d", "network": {"id": "c2b530e3-a253-4340-ba85-a76cf297211f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1520162603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61c731af2b5475f8d236b7d31b6073d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23817056-3d", "ovs_interfaceid": "23817056-3d15-4a89-8f18-84e4f7549efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.650833] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Acquired lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.651013] env[62558]: DEBUG nova.network.neutron [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Refreshing network info cache for port 23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.652675] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:80:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '23817056-3d15-4a89-8f18-84e4f7549efe', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.660577] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Creating folder: Project (d61c731af2b5475f8d236b7d31b6073d). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.662017] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44abc9fc-bd6a-4a42-a6b9-1c88d81a6a1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.672392] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Created folder: Project (d61c731af2b5475f8d236b7d31b6073d) in parent group-v272451. [ 792.672583] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Creating folder: Instances. Parent ref: group-v272481. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.672851] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-102b4001-b8ae-425e-82e6-1fa30ad69f57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.677597] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266589, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442578} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.678271] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.678468] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.678674] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e3dc8fc-320d-49d6-8a4d-0e7b1b0f720f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.684437] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 792.684437] env[62558]: value = "task-1266592" [ 792.684437] env[62558]: _type = "Task" [ 792.684437] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.686900] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Created folder: Instances in parent group-v272481. [ 792.687145] env[62558]: DEBUG oslo.service.loopingcall [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.690806] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.690934] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6529f3c-680a-4491-bdbd-4fe519ebbf1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.711769] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.713200] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.713200] env[62558]: value = "task-1266593" [ 792.713200] env[62558]: _type = "Task" [ 792.713200] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.721228] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266593, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.795478] env[62558]: DEBUG nova.scheduler.client.report [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.841650] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.866578] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.866835] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.867008] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.867218] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.867364] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.867511] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.867714] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.867872] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.868512] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.868765] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.868952] env[62558]: DEBUG nova.virt.hardware [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.869803] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e2c0ce-e90c-481b-a87a-a38f40caede9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.878413] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83625bf-a70c-444a-be1c-4b2d9963c259 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.205837] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063146} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.208394] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.209224] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35747356-ab87-4049-aac6-b7674a1345f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.242222] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.246587] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa214a52-b6c0-4c39-99c5-8dc0409d904e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.270257] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266593, 'name': CreateVM_Task, 'duration_secs': 0.291549} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.270439] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.272047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.273606] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.273954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.274763] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4cfccb9-30af-4995-9227-a8894361548b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.280283] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 793.280283] env[62558]: value = "task-1266594" [ 793.280283] env[62558]: _type = "Task" [ 793.280283] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.288766] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 793.288766] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521094d0-acf2-d01f-3da2-f4c032c4ad8a" [ 793.288766] env[62558]: _type = "Task" [ 793.288766] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.296686] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.302128] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.302698] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.306649] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521094d0-acf2-d01f-3da2-f4c032c4ad8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.307166] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.857s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.308669] env[62558]: INFO nova.compute.claims [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.365492] env[62558]: DEBUG nova.compute.manager [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Received event network-vif-plugged-c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.365735] env[62558]: DEBUG oslo_concurrency.lockutils [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] Acquiring lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.365934] env[62558]: DEBUG oslo_concurrency.lockutils [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.366123] env[62558]: DEBUG oslo_concurrency.lockutils [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.366328] env[62558]: DEBUG nova.compute.manager [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] No waiting events found dispatching network-vif-plugged-c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.366445] env[62558]: WARNING nova.compute.manager [req-676123bd-d27b-4a7f-8d3f-36c77f3b3aca req-97d3b4c0-69d8-4612-b439-74ea68579da8 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Received unexpected event network-vif-plugged-c8245937-784e-4e48-8246-3323fab7a3dc for instance with vm_state building and task_state spawning. [ 793.498012] env[62558]: DEBUG nova.network.neutron [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Updated VIF entry in instance network info cache for port 23817056-3d15-4a89-8f18-84e4f7549efe. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 793.498350] env[62558]: DEBUG nova.network.neutron [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Updating instance_info_cache with network_info: [{"id": "23817056-3d15-4a89-8f18-84e4f7549efe", "address": "fa:16:3e:83:80:3d", "network": {"id": "c2b530e3-a253-4340-ba85-a76cf297211f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1520162603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d61c731af2b5475f8d236b7d31b6073d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap23817056-3d", "ovs_interfaceid": "23817056-3d15-4a89-8f18-84e4f7549efe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.684319] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Successfully updated port: c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.791503] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266594, 'name': ReconfigVM_Task, 'duration_secs': 0.256984} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.795343] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.795684] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9dd2cad-3117-4bb3-8a5b-73c243b014de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.802791] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521094d0-acf2-d01f-3da2-f4c032c4ad8a, 'name': SearchDatastore_Task, 'duration_secs': 0.023244} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.804045] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.804286] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.804515] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.804657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.804828] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.805135] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 793.805135] env[62558]: value = "task-1266595" [ 793.805135] env[62558]: _type = "Task" [ 793.805135] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.805314] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8fdb59b6-148a-4dda-a4ab-5f1ff3e886af {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.808183] env[62558]: DEBUG nova.compute.utils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.813274] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.813361] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 793.819465] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.819640] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.823468] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12b3456d-8758-4974-898c-be9a76428467 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.826026] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266595, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.830109] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 793.830109] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523978e9-1a35-be9e-48af-b77b36f3ad4f" [ 793.830109] env[62558]: _type = "Task" [ 793.830109] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.838334] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523978e9-1a35-be9e-48af-b77b36f3ad4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.906137] env[62558]: DEBUG nova.policy [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '413d21eea08b4801af3bec013f1b8045', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b336ad28239649b7949dd21ad2df85bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.001540] env[62558]: DEBUG oslo_concurrency.lockutils [req-f0bd7754-f17a-4db5-bd62-7216e8b4ff08 req-bb0e3678-07f0-4269-8f35-1d1d8dbe8d91 service nova] Releasing lock "refresh_cache-8acd06fc-b040-468c-980f-6e17b2343c4a" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.188038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.188038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquired lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.188038] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.303574] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Successfully created port: 4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.313137] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.330418] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266595, 'name': Rename_Task, 'duration_secs': 0.137681} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.332799] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.336358] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43608dd6-31f5-470b-befb-dc501cd3e526 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.348336] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523978e9-1a35-be9e-48af-b77b36f3ad4f, 'name': SearchDatastore_Task, 'duration_secs': 0.009082} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.350914] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 794.350914] env[62558]: value = "task-1266596" [ 794.350914] env[62558]: _type = "Task" [ 794.350914] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.350914] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f763888a-c932-4636-addd-73674b7215b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.363044] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 794.363044] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eeb073-3ac3-35d8-4255-06bb4a0ba8c1" [ 794.363044] env[62558]: _type = "Task" [ 794.363044] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.369747] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266596, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.384754] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eeb073-3ac3-35d8-4255-06bb4a0ba8c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010052} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.385051] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.385519] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 8acd06fc-b040-468c-980f-6e17b2343c4a/8acd06fc-b040-468c-980f-6e17b2343c4a.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.385646] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-decb0078-3e81-4308-ad3a-b350fc6f26cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.394430] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 794.394430] env[62558]: value = "task-1266597" [ 794.394430] env[62558]: _type = "Task" [ 794.394430] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.402768] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.739551] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.794329] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15766f67-6f6c-4b01-ad8b-962bdd889028 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.805395] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50fb76c-d476-4ad2-8112-ad085a9695e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.844162] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6873ed4a-d42d-4456-afd6-3a2ff38931b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.859985] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45136213-c0e3-448d-8d17-e98ec7d52fa0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.873410] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266596, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.883027] env[62558]: DEBUG nova.compute.provider_tree [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.905876] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266597, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500907} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.906203] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 8acd06fc-b040-468c-980f-6e17b2343c4a/8acd06fc-b040-468c-980f-6e17b2343c4a.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 794.906414] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.906934] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6532458-71dd-4ab4-b817-bab82f6c8fc4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.913830] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 794.913830] env[62558]: value = "task-1266598" [ 794.913830] env[62558]: _type = "Task" [ 794.913830] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.921973] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.986686] env[62558]: DEBUG nova.network.neutron [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Updating instance_info_cache with network_info: [{"id": "c8245937-784e-4e48-8246-3323fab7a3dc", "address": "fa:16:3e:3a:1f:21", "network": {"id": "24b1931d-a5ae-47c2-bbc2-eba7045ca091", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2007897114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3625580997514184ab36f8258cf2358b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8245937-78", "ovs_interfaceid": "c8245937-784e-4e48-8246-3323fab7a3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.345448] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.364074] env[62558]: DEBUG oslo_vmware.api [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266596, 'name': PowerOnVM_Task, 'duration_secs': 0.582754} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.364345] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.364537] env[62558]: INFO nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Took 7.38 seconds to spawn the instance on the hypervisor. [ 795.364730] env[62558]: DEBUG nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.365535] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88b367a-dd0e-441b-8dce-8526cb5df113 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.382170] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.382170] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.382170] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.382492] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.382492] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.382492] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.382492] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.382492] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.382619] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.382619] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.382619] env[62558]: DEBUG nova.virt.hardware [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.382619] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2edf139-2486-459a-ad8d-d7b6a6621a31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.389561] env[62558]: DEBUG nova.scheduler.client.report [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.393747] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aada6b4-cc51-4e62-b26f-c39deba73858 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.425736] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095011} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.426666] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 795.427937] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372abb34-7c11-474a-a8eb-1a8025590cf6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.453133] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 8acd06fc-b040-468c-980f-6e17b2343c4a/8acd06fc-b040-468c-980f-6e17b2343c4a.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.453755] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef3b7120-eee2-430e-b623-25fb96908ebf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.473797] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 795.473797] env[62558]: value = "task-1266599" [ 795.473797] env[62558]: _type = "Task" [ 795.473797] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.481523] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.490145] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Releasing lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.490145] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Instance network_info: |[{"id": "c8245937-784e-4e48-8246-3323fab7a3dc", "address": "fa:16:3e:3a:1f:21", "network": {"id": "24b1931d-a5ae-47c2-bbc2-eba7045ca091", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2007897114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3625580997514184ab36f8258cf2358b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8245937-78", "ovs_interfaceid": "c8245937-784e-4e48-8246-3323fab7a3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 795.490492] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:1f:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13b62154-a0e1-4eed-bc30-6464b15993bb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8245937-784e-4e48-8246-3323fab7a3dc', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.497616] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Creating folder: Project (3625580997514184ab36f8258cf2358b). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.498025] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a15a5f3-6649-4eb5-9ae0-7e421412135d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.509115] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Created folder: Project (3625580997514184ab36f8258cf2358b) in parent group-v272451. [ 795.509565] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Creating folder: Instances. Parent ref: group-v272484. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.509824] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a77a802b-ef41-43cb-b063-c314d482a93a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.519734] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Created folder: Instances in parent group-v272484. [ 795.519734] env[62558]: DEBUG oslo.service.loopingcall [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.519734] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 795.520601] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f41d8bc-8cf6-44e1-bf7f-11f2fc8fd21b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.537844] env[62558]: DEBUG nova.compute.manager [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Received event network-changed-c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.537844] env[62558]: DEBUG nova.compute.manager [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Refreshing instance network info cache due to event network-changed-c8245937-784e-4e48-8246-3323fab7a3dc. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.537844] env[62558]: DEBUG oslo_concurrency.lockutils [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] Acquiring lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.537844] env[62558]: DEBUG oslo_concurrency.lockutils [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] Acquired lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.537844] env[62558]: DEBUG nova.network.neutron [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Refreshing network info cache for port c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.548833] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.548833] env[62558]: value = "task-1266602" [ 795.548833] env[62558]: _type = "Task" [ 795.548833] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.559080] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266602, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.901683] env[62558]: INFO nova.compute.manager [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Took 35.94 seconds to build instance. [ 795.903460] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.905361] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.907827] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.187s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.910043] env[62558]: INFO nova.compute.claims [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.982857] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266599, 'name': ReconfigVM_Task, 'duration_secs': 0.350014} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.983154] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 8acd06fc-b040-468c-980f-6e17b2343c4a/8acd06fc-b040-468c-980f-6e17b2343c4a.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.983757] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71939e81-c0ef-4940-ae47-543ded2eb6cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.989774] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 795.989774] env[62558]: value = "task-1266603" [ 795.989774] env[62558]: _type = "Task" [ 795.989774] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.997758] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266603, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.060225] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266602, 'name': CreateVM_Task, 'duration_secs': 0.418309} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.061851] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 796.061851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.061851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.061851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.062249] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10d121c1-76d1-4dce-9b28-4f097c184e8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.067673] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 796.067673] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5232695f-2d66-bc9a-d6d9-36719bf35b01" [ 796.067673] env[62558]: _type = "Task" [ 796.067673] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.075949] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5232695f-2d66-bc9a-d6d9-36719bf35b01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.142900] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Successfully updated port: 4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.295481] env[62558]: DEBUG nova.network.neutron [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Updated VIF entry in instance network info cache for port c8245937-784e-4e48-8246-3323fab7a3dc. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 796.296445] env[62558]: DEBUG nova.network.neutron [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Updating instance_info_cache with network_info: [{"id": "c8245937-784e-4e48-8246-3323fab7a3dc", "address": "fa:16:3e:3a:1f:21", "network": {"id": "24b1931d-a5ae-47c2-bbc2-eba7045ca091", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-2007897114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3625580997514184ab36f8258cf2358b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13b62154-a0e1-4eed-bc30-6464b15993bb", "external-id": "nsx-vlan-transportzone-514", "segmentation_id": 514, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8245937-78", "ovs_interfaceid": "c8245937-784e-4e48-8246-3323fab7a3dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.410699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c458525-3d92-4e74-b50f-e8d47f9fe82a tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.486s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.414165] env[62558]: DEBUG nova.compute.utils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.420499] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.420658] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.476986] env[62558]: DEBUG nova.policy [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c5503a5bf734483af68683a8c853a71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9d4de9f22ec414d90eb8c2ed9c42d18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.504114] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266603, 'name': Rename_Task, 'duration_secs': 0.174755} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.504591] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 796.504636] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75e4b7af-9e4e-492f-8a71-def28fa81bed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.511348] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 796.511348] env[62558]: value = "task-1266604" [ 796.511348] env[62558]: _type = "Task" [ 796.511348] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.519619] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266604, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.580768] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5232695f-2d66-bc9a-d6d9-36719bf35b01, 'name': SearchDatastore_Task, 'duration_secs': 0.029939} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.580768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.580768] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.580768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.580954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.580954] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.582578] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-929e0ed8-9ec8-48de-b12e-a28dac8e6847 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.590280] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.590496] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.591302] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3b98e61-c001-4dee-8917-6a042f3a36fc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.596961] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 796.596961] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a44c59-ff00-7569-3fc3-a7868beb4ab0" [ 796.596961] env[62558]: _type = "Task" [ 796.596961] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.612026] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a44c59-ff00-7569-3fc3-a7868beb4ab0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.620392] env[62558]: INFO nova.compute.manager [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Rescuing [ 796.620717] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.620891] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.621076] env[62558]: DEBUG nova.network.neutron [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.648293] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.648293] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquired lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.648293] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.781812] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Successfully created port: cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.801402] env[62558]: DEBUG oslo_concurrency.lockutils [req-901cb64f-b0c7-4a5e-904c-c6caf47a219e req-2663e2eb-7de3-4493-a62a-bb18f8aa61ab service nova] Releasing lock "refresh_cache-3c8c2d25-981e-49de-b5c0-e0d12dd91378" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.921221] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.928585] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.029818] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266604, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.110123] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a44c59-ff00-7569-3fc3-a7868beb4ab0, 'name': SearchDatastore_Task, 'duration_secs': 0.008552} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.115808] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fec53f5b-52fd-4fee-a5a9-024c803e22a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.119892] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 797.119892] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d07f43-99e1-7e78-1da3-d54d390160bd" [ 797.119892] env[62558]: _type = "Task" [ 797.119892] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.134096] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d07f43-99e1-7e78-1da3-d54d390160bd, 'name': SearchDatastore_Task, 'duration_secs': 0.011953} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.134791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.135102] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 3c8c2d25-981e-49de-b5c0-e0d12dd91378/3c8c2d25-981e-49de-b5c0-e0d12dd91378.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.135534] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a88b9cf-fd8c-4377-a887-d7bb62f1ae49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.145036] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 797.145036] env[62558]: value = "task-1266605" [ 797.145036] env[62558]: _type = "Task" [ 797.145036] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.161746] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.200266] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.235868] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "574525eb-0535-4664-8449-813c16e4781c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.236520] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.396232] env[62558]: DEBUG nova.network.neutron [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updating instance_info_cache with network_info: [{"id": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "address": "fa:16:3e:a9:64:2f", "network": {"id": "ad9fdca6-db0a-4237-8535-88789de4bc00", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-567677333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b336ad28239649b7949dd21ad2df85bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4867bc9b-01", "ovs_interfaceid": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.421467] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911bdf79-93c9-47ab-b4f2-6844f732dfee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.436367] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318a0426-2240-442b-be6f-e773a4399f6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.490930] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.491883] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13410888-f250-42ed-afbe-62bbca62753d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.503954] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061a7e02-a569-4229-ac94-b0fbf1f505eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.522445] env[62558]: DEBUG nova.compute.provider_tree [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.537793] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266604, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.539932] env[62558]: DEBUG nova.network.neutron [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [{"id": "15338668-abe5-414d-91c8-3396fdaf1b3e", "address": "fa:16:3e:f8:9c:60", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15338668-ab", "ovs_interfaceid": "15338668-abe5-414d-91c8-3396fdaf1b3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.564345] env[62558]: DEBUG nova.compute.manager [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Received event network-vif-plugged-4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.564726] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Acquiring lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.565230] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.565535] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.566076] env[62558]: DEBUG nova.compute.manager [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] No waiting events found dispatching network-vif-plugged-4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 797.566411] env[62558]: WARNING nova.compute.manager [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Received unexpected event network-vif-plugged-4867bc9b-0198-4b9c-ab0c-a41743956ee8 for instance with vm_state building and task_state spawning. [ 797.566684] env[62558]: DEBUG nova.compute.manager [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Received event network-changed-4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.566934] env[62558]: DEBUG nova.compute.manager [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Refreshing instance network info cache due to event network-changed-4867bc9b-0198-4b9c-ab0c-a41743956ee8. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 797.567223] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Acquiring lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.654676] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266605, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.901223] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Releasing lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.901471] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Instance network_info: |[{"id": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "address": "fa:16:3e:a9:64:2f", "network": {"id": "ad9fdca6-db0a-4237-8535-88789de4bc00", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-567677333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b336ad28239649b7949dd21ad2df85bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4867bc9b-01", "ovs_interfaceid": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.901788] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Acquired lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.901973] env[62558]: DEBUG nova.network.neutron [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Refreshing network info cache for port 4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 797.903349] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:64:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '748a5204-8f14-402c-9a6e-f3e6104db082', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4867bc9b-0198-4b9c-ab0c-a41743956ee8', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.910946] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Creating folder: Project (b336ad28239649b7949dd21ad2df85bd). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.911492] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6fbb3da2-6caa-4bf6-99ec-67e286d50a94 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.922570] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Created folder: Project (b336ad28239649b7949dd21ad2df85bd) in parent group-v272451. [ 797.922784] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Creating folder: Instances. Parent ref: group-v272487. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.923031] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96671f5b-1c19-410f-b17d-d4a682c5f666 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.935108] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Created folder: Instances in parent group-v272487. [ 797.935349] env[62558]: DEBUG oslo.service.loopingcall [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.935572] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.935811] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f91e64c3-61b9-4810-9655-cfc811c1897e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.950346] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.957615] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.957615] env[62558]: value = "task-1266608" [ 797.957615] env[62558]: _type = "Task" [ 797.957615] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.965810] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266608, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.975802] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.975892] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.976149] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.976319] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.976440] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.976588] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.977504] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.977595] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.977759] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.977929] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.978116] env[62558]: DEBUG nova.virt.hardware [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.978928] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27174e20-e916-4ba9-85f0-f3f057d877a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.988952] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a79bfb-20e7-44bd-beee-33c1a0ceb7ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.030395] env[62558]: DEBUG nova.scheduler.client.report [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.033739] env[62558]: DEBUG oslo_vmware.api [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266604, 'name': PowerOnVM_Task, 'duration_secs': 1.225211} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.033989] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 798.034206] env[62558]: INFO nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 7.68 seconds to spawn the instance on the hypervisor. [ 798.034452] env[62558]: DEBUG nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.035461] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50ea735-bcce-41e0-9608-8f50bf2ba398 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.044122] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.154669] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266605, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.209151] env[62558]: DEBUG nova.compute.manager [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Received event network-vif-plugged-cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.209151] env[62558]: DEBUG oslo_concurrency.lockutils [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] Acquiring lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.209151] env[62558]: DEBUG oslo_concurrency.lockutils [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] Lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.209151] env[62558]: DEBUG oslo_concurrency.lockutils [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] Lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.209939] env[62558]: DEBUG nova.compute.manager [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] No waiting events found dispatching network-vif-plugged-cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 798.210279] env[62558]: WARNING nova.compute.manager [req-27562c14-7434-49a3-b530-74c129fd509b req-357514d9-161c-49bb-a9bf-8f81e8360396 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Received unexpected event network-vif-plugged-cc1b2699-3507-45b0-afe0-572fd6966fdf for instance with vm_state building and task_state spawning. [ 798.303345] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Successfully updated port: cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.467020] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266608, 'name': CreateVM_Task, 'duration_secs': 0.328973} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.467306] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.468277] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.468277] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.468453] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.469036] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1c4891-469d-4b83-bffd-bbb813987697 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.473458] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 798.473458] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5235c3b2-60fb-31c0-f144-ba8d1bd7894f" [ 798.473458] env[62558]: _type = "Task" [ 798.473458] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.481257] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5235c3b2-60fb-31c0-f144-ba8d1bd7894f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.536599] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.537111] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.539880] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.295s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.541193] env[62558]: INFO nova.compute.claims [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.555844] env[62558]: INFO nova.compute.manager [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 37.44 seconds to build instance. [ 798.572654] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.574052] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95c62238-6c0f-4316-a897-a2f7a56af884 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.581396] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 798.581396] env[62558]: value = "task-1266609" [ 798.581396] env[62558]: _type = "Task" [ 798.581396] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.594473] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266609, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.622255] env[62558]: DEBUG nova.network.neutron [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updated VIF entry in instance network info cache for port 4867bc9b-0198-4b9c-ab0c-a41743956ee8. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 798.622657] env[62558]: DEBUG nova.network.neutron [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updating instance_info_cache with network_info: [{"id": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "address": "fa:16:3e:a9:64:2f", "network": {"id": "ad9fdca6-db0a-4237-8535-88789de4bc00", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-567677333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b336ad28239649b7949dd21ad2df85bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4867bc9b-01", "ovs_interfaceid": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.655029] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266605, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.473529} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.655189] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 3c8c2d25-981e-49de-b5c0-e0d12dd91378/3c8c2d25-981e-49de-b5c0-e0d12dd91378.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.655490] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.655777] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-718cc54d-56c3-456f-98d0-520f99080d36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.661595] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 798.661595] env[62558]: value = "task-1266610" [ 798.661595] env[62558]: _type = "Task" [ 798.661595] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.669535] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.806931] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.806931] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.806931] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.984101] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5235c3b2-60fb-31c0-f144-ba8d1bd7894f, 'name': SearchDatastore_Task, 'duration_secs': 0.008787} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.984449] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.984703] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.984940] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.985106] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.985310] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.985590] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fd47df4-f3ef-4dab-aa1f-4072667f2448 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.994263] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.994508] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.995265] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05eec9f2-5ac9-488e-bb45-332fc63aeba0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.001144] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 799.001144] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b0bab9-be59-d3e6-a549-b09c3178caac" [ 799.001144] env[62558]: _type = "Task" [ 799.001144] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.008626] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b0bab9-be59-d3e6-a549-b09c3178caac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.045579] env[62558]: DEBUG nova.compute.utils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.048788] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.048927] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.057181] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6523958a-15de-421d-bd4a-6502540121de tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.297s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.095606] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266609, 'name': PowerOffVM_Task, 'duration_secs': 0.459252} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.098702] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 799.098702] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66df656-d716-4766-a21c-d3b51f0e1419 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.102886] env[62558]: DEBUG nova.policy [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3faaab5e759446789a31b57b0abe52e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebd5df2c8e5f42f5b4b3da70f8793ed9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.118784] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfeb62c-73f2-4bcd-a3e8-43ef763d5301 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.126642] env[62558]: DEBUG oslo_concurrency.lockutils [req-217ddc53-d9ba-4396-bb37-6313072ddddc req-14b2267b-1259-45d1-ac84-b8317f08a6b9 service nova] Releasing lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.152352] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.152543] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1790cacc-6d8b-437a-a6dd-d4feecd4e44e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.159404] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 799.159404] env[62558]: value = "task-1266611" [ 799.159404] env[62558]: _type = "Task" [ 799.159404] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.174864] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064822} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.175221] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.175535] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 799.175720] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.175952] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.176110] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.176289] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.177288] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a10421-2a1b-47a5-a810-3fc636bf4aaf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.179652] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5eea93d1-4568-4f3c-90ca-8b576632c982 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.201725] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 3c8c2d25-981e-49de-b5c0-e0d12dd91378/3c8c2d25-981e-49de-b5c0-e0d12dd91378.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.203900] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a0f5b21-0685-4092-b17a-6d98dc7a446c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.217477] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.217657] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.218425] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ab8e23-02ad-4e0b-a008-1300a7bc2c3e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.224572] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 799.224572] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d970c7-9765-4e44-b79b-fa4d43d38ea0" [ 799.224572] env[62558]: _type = "Task" [ 799.224572] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.228451] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 799.228451] env[62558]: value = "task-1266612" [ 799.228451] env[62558]: _type = "Task" [ 799.228451] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.234619] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d970c7-9765-4e44-b79b-fa4d43d38ea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.239684] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266612, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.346040] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.421211] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Successfully created port: bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.514757] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b0bab9-be59-d3e6-a549-b09c3178caac, 'name': SearchDatastore_Task, 'duration_secs': 0.008951} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.515936] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4cc9d57-0e30-42ea-ba14-d7a534764f5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.521060] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 799.521060] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1f1f3-d111-b93a-14dd-310f061176b7" [ 799.521060] env[62558]: _type = "Task" [ 799.521060] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.525536] env[62558]: DEBUG nova.network.neutron [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Updating instance_info_cache with network_info: [{"id": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "address": "fa:16:3e:68:2e:90", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1b2699-35", "ovs_interfaceid": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.532521] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1f1f3-d111-b93a-14dd-310f061176b7, 'name': SearchDatastore_Task, 'duration_secs': 0.008577} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.532761] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.533716] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a8fef137-a651-4724-8ad1-a3e6ddcb5b03/a8fef137-a651-4724-8ad1-a3e6ddcb5b03.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.533716] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c068c75d-5fcb-42bf-a8f0-c78618aede39 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.539301] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 799.539301] env[62558]: value = "task-1266613" [ 799.539301] env[62558]: _type = "Task" [ 799.539301] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.548305] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.549876] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.560039] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.739790] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d970c7-9765-4e44-b79b-fa4d43d38ea0, 'name': SearchDatastore_Task, 'duration_secs': 0.009959} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.741245] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b87a18d-1dc3-4cde-8225-0a64349cf2df {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.751519] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266612, 'name': ReconfigVM_Task, 'duration_secs': 0.311518} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.752172] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 3c8c2d25-981e-49de-b5c0-e0d12dd91378/3c8c2d25-981e-49de-b5c0-e0d12dd91378.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.753132] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9cc4e047-0be3-4c5a-b035-6ff49293c6d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.757149] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 799.757149] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520a0597-8374-9179-e94d-0b48276fb6c0" [ 799.757149] env[62558]: _type = "Task" [ 799.757149] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.765453] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 799.765453] env[62558]: value = "task-1266614" [ 799.765453] env[62558]: _type = "Task" [ 799.765453] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.773043] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520a0597-8374-9179-e94d-0b48276fb6c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.782378] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266614, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.988643] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca5e5c2-d513-4d5a-9a0f-1d8d4672cbcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.998444] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe40a26d-0d8f-4fcd-a850-af6d46b9258b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.031284] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.031887] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance network_info: |[{"id": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "address": "fa:16:3e:68:2e:90", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1b2699-35", "ovs_interfaceid": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.033019] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:2e:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f256cfee-512d-4192-9aca-6750fdb1cd4c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc1b2699-3507-45b0-afe0-572fd6966fdf', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.042023] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating folder: Project (f9d4de9f22ec414d90eb8c2ed9c42d18). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.043590] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5079c1db-56d1-491e-ad80-38daa5564355 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.046593] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ad19cd6-eac0-4408-acd3-7d156c3af9f8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.061217] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266613, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514685} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.065150] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a8fef137-a651-4724-8ad1-a3e6ddcb5b03/a8fef137-a651-4724-8ad1-a3e6ddcb5b03.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 800.065389] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.067768] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created folder: Project (f9d4de9f22ec414d90eb8c2ed9c42d18) in parent group-v272451. [ 800.067949] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating folder: Instances. Parent ref: group-v272490. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.069996] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb1f984c-7a0f-4989-9978-f4fe4dd50538 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.074833] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2c404f-f7eb-4d2c-8c36-e935e110e5d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.077659] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28529654-c1bf-46b4-a265-827833f44b1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.092367] env[62558]: DEBUG nova.compute.provider_tree [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.099017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.099017] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created folder: Instances in parent group-v272490. [ 800.099017] env[62558]: DEBUG oslo.service.loopingcall [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.099017] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 800.099017] env[62558]: value = "task-1266616" [ 800.099017] env[62558]: _type = "Task" [ 800.099017] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.099017] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.099366] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f2dc374-d3c0-443e-8b68-1b826d4092db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.123708] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.125189] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.125189] env[62558]: value = "task-1266618" [ 800.125189] env[62558]: _type = "Task" [ 800.125189] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.133027] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266618, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.235877] env[62558]: DEBUG nova.compute.manager [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Received event network-changed-cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.235877] env[62558]: DEBUG nova.compute.manager [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Refreshing instance network info cache due to event network-changed-cc1b2699-3507-45b0-afe0-572fd6966fdf. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.235877] env[62558]: DEBUG oslo_concurrency.lockutils [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] Acquiring lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.235877] env[62558]: DEBUG oslo_concurrency.lockutils [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] Acquired lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.235877] env[62558]: DEBUG nova.network.neutron [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Refreshing network info cache for port cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.254392] env[62558]: DEBUG nova.compute.manager [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 800.255121] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649f7024-cf52-49d3-86be-7d934cd21c2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.279349] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520a0597-8374-9179-e94d-0b48276fb6c0, 'name': SearchDatastore_Task, 'duration_secs': 0.015042} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.284317] env[62558]: DEBUG oslo_concurrency.lockutils [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.284837] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. {{(pid=62558) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 800.285232] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266614, 'name': Rename_Task, 'duration_secs': 0.381277} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.285540] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dea5f0f3-1b1d-42cd-8b9e-851f654d8839 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.288731] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.289477] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63ca882e-c05e-4580-b61c-9f16e1a105e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.296654] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 800.296654] env[62558]: value = "task-1266619" [ 800.296654] env[62558]: _type = "Task" [ 800.296654] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.299327] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 800.299327] env[62558]: value = "task-1266620" [ 800.299327] env[62558]: _type = "Task" [ 800.299327] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.315319] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.318765] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266620, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.566329] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.592701] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.593014] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.593362] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.593634] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.593777] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.593956] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.594237] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.594472] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.594748] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.594948] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.595169] env[62558]: DEBUG nova.virt.hardware [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.596247] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dff03f-f4d4-4664-b14c-2fce0aa194b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.600271] env[62558]: DEBUG nova.scheduler.client.report [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.618679] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fadd8cb-0b00-4f0b-8b1f-b8342729817b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.623188] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093826} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.624056] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.625366] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde588b8-c672-4942-b8c7-80b01091acca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.662978] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a8fef137-a651-4724-8ad1-a3e6ddcb5b03/a8fef137-a651-4724-8ad1-a3e6ddcb5b03.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.668030] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cea07bc0-7b05-46be-a9eb-c65d3f91ccf1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.683135] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266618, 'name': CreateVM_Task, 'duration_secs': 0.345344} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.683261] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 800.684468] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.684699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.685063] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.685699] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e9594f9-e329-4520-bf30-da910f92ba93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.690745] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 800.690745] env[62558]: value = "task-1266621" [ 800.690745] env[62558]: _type = "Task" [ 800.690745] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.695645] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 800.695645] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a402e-1d9f-5ee8-7094-6f61c75d0049" [ 800.695645] env[62558]: _type = "Task" [ 800.695645] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.703788] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266621, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.712666] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a402e-1d9f-5ee8-7094-6f61c75d0049, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.766324] env[62558]: INFO nova.compute.manager [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] instance snapshotting [ 800.769127] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94341898-eca4-4f4d-9403-8eab5e0daddf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.794250] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1df456-9392-4ca7-95c6-1ca8ac0da420 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.818504] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266619, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.820599] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266620, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.998071] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Successfully updated port: bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.110251] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.110766] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.114206] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.553s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.115666] env[62558]: INFO nova.compute.claims [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.150552] env[62558]: DEBUG nova.network.neutron [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Updated VIF entry in instance network info cache for port cc1b2699-3507-45b0-afe0-572fd6966fdf. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.150552] env[62558]: DEBUG nova.network.neutron [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Updating instance_info_cache with network_info: [{"id": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "address": "fa:16:3e:68:2e:90", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc1b2699-35", "ovs_interfaceid": "cc1b2699-3507-45b0-afe0-572fd6966fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.204963] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.209808] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a402e-1d9f-5ee8-7094-6f61c75d0049, 'name': SearchDatastore_Task, 'duration_secs': 0.025342} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.210145] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.210599] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.210672] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.210881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.210964] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.211260] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50081211-c542-42e6-926e-d23ac4dd0a65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.229031] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.229228] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.230095] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95d848a8-3e42-4430-bcdc-73b22977c7bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.235642] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 801.235642] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f47adb-0840-e413-f80a-a47b059a02cb" [ 801.235642] env[62558]: _type = "Task" [ 801.235642] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.244857] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f47adb-0840-e413-f80a-a47b059a02cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.319047] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 801.319047] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266619, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561587} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.319047] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7e9e485b-8d0e-4e96-9cb3-b6bc5f9c824a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.322751] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. [ 801.323579] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d159beb-a254-469e-94d3-235ed065fe77 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.330973] env[62558]: DEBUG oslo_vmware.api [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266620, 'name': PowerOnVM_Task, 'duration_secs': 0.6652} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.345356] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.345669] env[62558]: INFO nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Took 8.50 seconds to spawn the instance on the hypervisor. [ 801.345897] env[62558]: DEBUG nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.346299] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 801.346299] env[62558]: value = "task-1266622" [ 801.346299] env[62558]: _type = "Task" [ 801.346299] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.355566] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.356443] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd94cd3-d9d5-4418-a013-9f87bc8a0087 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.359186] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de77f1c1-e65d-48e1-bfd4-7a2ec10bc1eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.386333] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266622, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.386627] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 801.386627] env[62558]: value = "task-1266623" [ 801.386627] env[62558]: _type = "Task" [ 801.386627] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.394498] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266623, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.501281] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.501487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.501665] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.620517] env[62558]: DEBUG nova.compute.utils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.624357] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.624541] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.653863] env[62558]: DEBUG oslo_concurrency.lockutils [req-24d33c9f-3b23-45c0-8581-784c4b349e90 req-b6b6f05c-ca56-4f33-bc0e-4f068dc8ae91 service nova] Releasing lock "refresh_cache-6d89107f-0727-4b8c-863d-d1e635000bff" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.663337] env[62558]: DEBUG nova.policy [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da1f27f9a6ac44a0a2dc286b7f0888a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fabbd2127c2445aa96cf0d513541d32', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.702330] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.746690] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f47adb-0840-e413-f80a-a47b059a02cb, 'name': SearchDatastore_Task, 'duration_secs': 0.079002} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.747311] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021f1b87-6e78-4060-b315-3e1679deb5da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.756918] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 801.756918] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c12bfb-72b0-7f6f-65dd-f9de3b7080e8" [ 801.756918] env[62558]: _type = "Task" [ 801.756918] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.764901] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c12bfb-72b0-7f6f-65dd-f9de3b7080e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.869408] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266622, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.901455] env[62558]: INFO nova.compute.manager [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Took 37.43 seconds to build instance. [ 801.909540] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.048354] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.052595] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Successfully created port: 2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.127633] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.205889] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266621, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.206499] env[62558]: DEBUG nova.network.neutron [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.274915] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c12bfb-72b0-7f6f-65dd-f9de3b7080e8, 'name': SearchDatastore_Task, 'duration_secs': 0.048569} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.277768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.278114] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.279310] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b0e5fc4-9b32-4499-8746-953b3cb88cd1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.285440] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 802.285440] env[62558]: value = "task-1266624" [ 802.285440] env[62558]: _type = "Task" [ 802.285440] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.296597] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.353414] env[62558]: DEBUG nova.compute.manager [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Received event network-vif-plugged-bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.353653] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.353870] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.354276] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.354276] env[62558]: DEBUG nova.compute.manager [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] No waiting events found dispatching network-vif-plugged-bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.354384] env[62558]: WARNING nova.compute.manager [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Received unexpected event network-vif-plugged-bdbc998f-cabe-438b-b3ce-6477b78a13e7 for instance with vm_state building and task_state spawning. [ 802.354545] env[62558]: DEBUG nova.compute.manager [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Received event network-changed-bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.354713] env[62558]: DEBUG nova.compute.manager [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Refreshing instance network info cache due to event network-changed-bdbc998f-cabe-438b-b3ce-6477b78a13e7. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.354873] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Acquiring lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.374980] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266622, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.405862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7699a9e0-bfc9-497d-84c6-d8e9ad228425 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.160s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.406167] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266623, 'name': ReconfigVM_Task, 'duration_secs': 0.843108} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.409863] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.411288] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f6d1be-5de8-4e02-8117-3f0afca6a305 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.445382] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3de24491-b0df-46a2-a62d-e22097859864 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.461092] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 802.461092] env[62558]: value = "task-1266625" [ 802.461092] env[62558]: _type = "Task" [ 802.461092] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.470978] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.564414] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266e06e1-69e0-45fc-ba4e-e9200b363143 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.573941] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f61253-2117-4fa2-93c8-5ed05a6ede2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.609083] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2618c13-d485-4370-9600-484bbdd9b7fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.618956] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7a7776-884a-4a1a-9698-f0cc9b66f55a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.639917] env[62558]: DEBUG nova.compute.provider_tree [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.706311] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266621, 'name': ReconfigVM_Task, 'duration_secs': 1.527897} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.708034] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a8fef137-a651-4724-8ad1-a3e6ddcb5b03/a8fef137-a651-4724-8ad1-a3e6ddcb5b03.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.709298] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.709615] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Instance network_info: |[{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 802.710205] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b26e755c-0b6e-4c36-aec8-99610a1df0b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.712330] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Acquired lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.712524] env[62558]: DEBUG nova.network.neutron [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Refreshing network info cache for port bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.713913] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:00:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53915f38-d7a0-42ec-8b30-1eacfb2cc379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bdbc998f-cabe-438b-b3ce-6477b78a13e7', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.722587] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Creating folder: Project (ebd5df2c8e5f42f5b4b3da70f8793ed9). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.726311] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7885205-1c64-4e20-9cfb-4a884bda194e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.737743] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 802.737743] env[62558]: value = "task-1266626" [ 802.737743] env[62558]: _type = "Task" [ 802.737743] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.742820] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Created folder: Project (ebd5df2c8e5f42f5b4b3da70f8793ed9) in parent group-v272451. [ 802.743021] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Creating folder: Instances. Parent ref: group-v272494. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.743705] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab2b7a63-454d-4d08-83fb-0d66dde09206 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.749548] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266626, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.759713] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Created folder: Instances in parent group-v272494. [ 802.760017] env[62558]: DEBUG oslo.service.loopingcall [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.760258] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.760601] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25faab96-ea8d-44ed-90e1-4879868e8c51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.787415] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.787415] env[62558]: value = "task-1266629" [ 802.787415] env[62558]: _type = "Task" [ 802.787415] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.805516] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266629, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.805924] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266624, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.872926] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266622, 'name': CreateSnapshot_Task, 'duration_secs': 1.116012} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.873214] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 802.874019] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0ea11d-ffa4-4357-9575-96ace7cef0e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.911295] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.971272] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266625, 'name': ReconfigVM_Task, 'duration_secs': 0.172073} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.971272] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.971272] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5ee5870-fd5d-40aa-9934-3391017388a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.977545] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 802.977545] env[62558]: value = "task-1266630" [ 802.977545] env[62558]: _type = "Task" [ 802.977545] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.987139] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.997111] env[62558]: DEBUG nova.network.neutron [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updated VIF entry in instance network info cache for port bdbc998f-cabe-438b-b3ce-6477b78a13e7. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.997679] env[62558]: DEBUG nova.network.neutron [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.142850] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.146048] env[62558]: DEBUG nova.scheduler.client.report [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.169159] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.169425] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.169578] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.169756] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.169899] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.170056] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.170264] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.170427] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.170608] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.170771] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.170944] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.172032] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9688f99-1dd3-41a2-8732-c9048242062f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.180628] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9700aca5-d354-4ee3-9d15-b89daa2f0bc9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.247173] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266626, 'name': Rename_Task, 'duration_secs': 0.173665} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.247442] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.247685] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a28c3fc6-d63e-47f7-985f-83abe1392fbb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.254482] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 803.254482] env[62558]: value = "task-1266631" [ 803.254482] env[62558]: _type = "Task" [ 803.254482] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.262167] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.298703] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54735} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.299285] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.299524] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.299771] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aec2f47b-9c58-4cdc-a11b-2e80f628e16c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.304371] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266629, 'name': CreateVM_Task, 'duration_secs': 0.33034} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.304822] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 803.305504] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.305685] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.305988] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 803.306229] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1db4f19-610a-4594-a57c-eaf0fbe4bdfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.308887] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 803.308887] env[62558]: value = "task-1266632" [ 803.308887] env[62558]: _type = "Task" [ 803.308887] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.312975] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 803.312975] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52875fa7-e929-85d3-4411-e0ad3b4880f0" [ 803.312975] env[62558]: _type = "Task" [ 803.312975] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.320971] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266632, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.325343] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52875fa7-e929-85d3-4411-e0ad3b4880f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.392289] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 803.392635] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-514e02be-68fc-4fbf-b39d-e7a974a39e01 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.402203] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 803.402203] env[62558]: value = "task-1266633" [ 803.402203] env[62558]: _type = "Task" [ 803.402203] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.411841] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266633, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.434068] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.490013] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266630, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.500958] env[62558]: DEBUG oslo_concurrency.lockutils [req-2edc123d-bf30-4542-8e8a-4f8c275a6661 req-c3c7072c-92b9-4af9-b329-1e8adf69a3b1 service nova] Releasing lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.650958] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.651841] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.656593] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.998s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.659255] env[62558]: INFO nova.compute.claims [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.769509] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266631, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.820032] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266632, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141079} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.820956] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.821417] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414daadb-a668-4e4e-817d-9ff6279a5928 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.827659] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52875fa7-e929-85d3-4411-e0ad3b4880f0, 'name': SearchDatastore_Task, 'duration_secs': 0.026625} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.828297] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.828621] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.828878] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.829042] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.829231] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.829485] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a08ad92-15da-46e9-97c2-f15424f8848c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.851072] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.851984] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce2c3777-a073-4db0-b12b-0f81560f1a31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.870561] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.872496] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.873116] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c93f765-861a-48a0-8640-c4ec55f11f0a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.877108] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 803.877108] env[62558]: value = "task-1266634" [ 803.877108] env[62558]: _type = "Task" [ 803.877108] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.878689] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 803.878689] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bad207-678e-243f-3efe-31842a36f441" [ 803.878689] env[62558]: _type = "Task" [ 803.878689] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.891484] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bad207-678e-243f-3efe-31842a36f441, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.896811] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266634, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.915622] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266633, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.924605] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Successfully updated port: 2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.931698] env[62558]: DEBUG nova.compute.manager [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Received event network-vif-plugged-2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.932036] env[62558]: DEBUG oslo_concurrency.lockutils [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] Acquiring lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.932297] env[62558]: DEBUG oslo_concurrency.lockutils [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.932521] env[62558]: DEBUG oslo_concurrency.lockutils [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.932710] env[62558]: DEBUG nova.compute.manager [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] No waiting events found dispatching network-vif-plugged-2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.933963] env[62558]: WARNING nova.compute.manager [req-feb92328-cca2-4b6b-8034-ddb3704e92fd req-a23de8cf-da7f-471a-a465-21c60dbff933 service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Received unexpected event network-vif-plugged-2f4de907-3732-4ed6-a741-08fa0ce7db8f for instance with vm_state building and task_state spawning. [ 803.989996] env[62558]: DEBUG oslo_vmware.api [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266630, 'name': PowerOnVM_Task, 'duration_secs': 0.93455} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.990261] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.993263] env[62558]: DEBUG nova.compute.manager [None req-076fea0a-07b3-4e78-97d5-d9b841d7e6a5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.994054] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97357e5-6cef-461e-80c6-746cc11c4fe1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.169677] env[62558]: DEBUG nova.compute.utils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.175744] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.175930] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 804.264854] env[62558]: DEBUG oslo_vmware.api [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266631, 'name': PowerOnVM_Task, 'duration_secs': 0.533703} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.265152] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.265410] env[62558]: INFO nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Took 8.92 seconds to spawn the instance on the hypervisor. [ 804.265535] env[62558]: DEBUG nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.266336] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2566f3-2ae3-4a2b-9c56-25d4bbbec1d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.271069] env[62558]: DEBUG nova.policy [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da1f27f9a6ac44a0a2dc286b7f0888a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fabbd2127c2445aa96cf0d513541d32', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.392582] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266634, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.396112] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bad207-678e-243f-3efe-31842a36f441, 'name': SearchDatastore_Task, 'duration_secs': 0.020675} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.396883] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-439fe11e-2343-4805-9b56-6d09d4541a30 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.402249] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 804.402249] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a6814c-fc6e-a1a6-6d8f-10891b0693a0" [ 804.402249] env[62558]: _type = "Task" [ 804.402249] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.413307] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a6814c-fc6e-a1a6-6d8f-10891b0693a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.416938] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266633, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.430663] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.430874] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.434075] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.676727] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.683174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.683557] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.683873] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.684178] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.684444] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.686505] env[62558]: INFO nova.compute.manager [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Terminating instance [ 804.695019] env[62558]: DEBUG nova.compute.manager [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.695019] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.695019] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2d44cb-efd2-4110-b71c-689f85d70c07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.702129] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.702545] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55674a0f-73ba-4ea3-9374-1e88d29b3ea3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.709907] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 804.709907] env[62558]: value = "task-1266635" [ 804.709907] env[62558]: _type = "Task" [ 804.709907] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.724037] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.788217] env[62558]: INFO nova.compute.manager [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Took 38.31 seconds to build instance. [ 804.802238] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Successfully created port: e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.888685] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266634, 'name': ReconfigVM_Task, 'duration_secs': 0.608811} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.889055] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.892325] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca7db79c-8c62-419e-bfba-08fa8eee4c91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.899261] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 804.899261] env[62558]: value = "task-1266636" [ 804.899261] env[62558]: _type = "Task" [ 804.899261] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.918363] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266636, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.925952] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a6814c-fc6e-a1a6-6d8f-10891b0693a0, 'name': SearchDatastore_Task, 'duration_secs': 0.010139} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.931432] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.931707] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] a1d242e6-1561-4bd4-8e39-281ab6346661/a1d242e6-1561-4bd4-8e39-281ab6346661.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.931987] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266633, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.932469] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-146d9d6a-9439-4704-a2c5-31224474dd87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.938296] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 804.938296] env[62558]: value = "task-1266637" [ 804.938296] env[62558]: _type = "Task" [ 804.938296] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.947330] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.978913] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.123232] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b968e083-0f2d-47cb-8595-828e46743242 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.136255] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Updating instance_info_cache with network_info: [{"id": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "address": "fa:16:3e:47:82:5f", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4de907-37", "ovs_interfaceid": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.138964] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f2d2bf-257b-45cf-b8a8-2449cac3ced5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.171368] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5e7780-222d-4a1e-89ed-550e757be1e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.180016] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcf945c-6ecf-45c9-bfc5-9468c6116d0d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.205408] env[62558]: DEBUG nova.compute.provider_tree [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.227994] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266635, 'name': PowerOffVM_Task, 'duration_secs': 0.161455} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.228270] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.228450] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.228737] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b59a2073-391f-44d9-b4fa-4f4442aa9ebf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.291881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-724f5d24-31cf-4bda-9d39-8b0193d52f66 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.184s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.302025] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.302025] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.302025] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Deleting the datastore file [datastore2] 3c8c2d25-981e-49de-b5c0-e0d12dd91378 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.302025] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22f6e289-3f54-4867-a6c8-7a7a6101a65e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.308580] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for the task: (returnval){ [ 805.308580] env[62558]: value = "task-1266639" [ 805.308580] env[62558]: _type = "Task" [ 805.308580] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.321846] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.410398] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266636, 'name': Rename_Task, 'duration_secs': 0.145593} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.413906] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.414849] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc5f7353-3e47-4d21-8daf-92fbb2640cbb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.422026] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266633, 'name': CloneVM_Task, 'duration_secs': 1.72057} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.423469] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Created linked-clone VM from snapshot [ 805.423820] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 805.423820] env[62558]: value = "task-1266640" [ 805.423820] env[62558]: _type = "Task" [ 805.423820] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.424582] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a976dc75-45ab-41ff-8d33-7caf6181c78b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.437350] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266640, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.443506] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Uploading image 3d2a9cd1-658b-4ac6-b24b-40043736ee7b {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 805.454745] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266637, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.477793] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 805.477793] env[62558]: value = "vm-272497" [ 805.477793] env[62558]: _type = "VirtualMachine" [ 805.477793] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 805.478285] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-df1038f2-030d-49ad-a866-155b87646472 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.489185] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lease: (returnval){ [ 805.489185] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521c1c5b-eeb5-f69f-d7f5-9c3d9d8e9ebb" [ 805.489185] env[62558]: _type = "HttpNfcLease" [ 805.489185] env[62558]: } obtained for exporting VM: (result){ [ 805.489185] env[62558]: value = "vm-272497" [ 805.489185] env[62558]: _type = "VirtualMachine" [ 805.489185] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 805.489493] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the lease: (returnval){ [ 805.489493] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521c1c5b-eeb5-f69f-d7f5-9c3d9d8e9ebb" [ 805.489493] env[62558]: _type = "HttpNfcLease" [ 805.489493] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 805.498718] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 805.498718] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521c1c5b-eeb5-f69f-d7f5-9c3d9d8e9ebb" [ 805.498718] env[62558]: _type = "HttpNfcLease" [ 805.498718] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 805.644240] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.644713] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Instance network_info: |[{"id": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "address": "fa:16:3e:47:82:5f", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4de907-37", "ovs_interfaceid": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.645017] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:82:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f4de907-3732-4ed6-a741-08fa0ce7db8f', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.653058] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Creating folder: Project (9fabbd2127c2445aa96cf0d513541d32). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.654185] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67fe4764-ab88-4237-a6a1-021cd0a8df22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.664085] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Created folder: Project (9fabbd2127c2445aa96cf0d513541d32) in parent group-v272451. [ 805.664296] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Creating folder: Instances. Parent ref: group-v272498. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.664573] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-240c8233-29b6-443b-b1ae-ce4d0770fe5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.674039] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Created folder: Instances in parent group-v272498. [ 805.674039] env[62558]: DEBUG oslo.service.loopingcall [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.674039] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.674039] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d125936-b66f-489b-9a63-fc279ee4d7da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.692744] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.692744] env[62558]: value = "task-1266644" [ 805.692744] env[62558]: _type = "Task" [ 805.692744] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.696589] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.703179] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266644, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.721872] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.722219] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.722394] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.722587] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.722733] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.722901] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.723167] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.723327] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.723701] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.723701] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.723905] env[62558]: DEBUG nova.virt.hardware [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.724798] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a7a2b5-15d5-4288-b988-46ac92db9849 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.732728] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd493dae-90b0-4af2-9f1e-5dbeb06c406c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.739154] env[62558]: ERROR nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [req-5634381d-4e5f-442d-b901-3151c69d0539] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5634381d-4e5f-442d-b901-3151c69d0539"}]} [ 805.761521] env[62558]: DEBUG nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 805.783719] env[62558]: DEBUG nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 805.784017] env[62558]: DEBUG nova.compute.provider_tree [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.796968] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.800869] env[62558]: DEBUG nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 805.820057] env[62558]: DEBUG oslo_vmware.api [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Task: {'id': task-1266639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312026} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.820057] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.820057] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.820274] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.820309] env[62558]: INFO nova.compute.manager [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Took 1.13 seconds to destroy the instance on the hypervisor. [ 805.820555] env[62558]: DEBUG oslo.service.loopingcall [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.820770] env[62558]: DEBUG nova.compute.manager [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.820943] env[62558]: DEBUG nova.network.neutron [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 805.825783] env[62558]: DEBUG nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 805.938211] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266640, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.953664] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628715} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.956874] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] a1d242e6-1561-4bd4-8e39-281ab6346661/a1d242e6-1561-4bd4-8e39-281ab6346661.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.957198] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.957741] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c53c3feb-a235-42cf-a758-707d4de4e324 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.965405] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 805.965405] env[62558]: value = "task-1266645" [ 805.965405] env[62558]: _type = "Task" [ 805.965405] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.977112] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.988388] env[62558]: DEBUG nova.compute.manager [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Received event network-changed-2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.988727] env[62558]: DEBUG nova.compute.manager [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Refreshing instance network info cache due to event network-changed-2f4de907-3732-4ed6-a741-08fa0ce7db8f. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.988987] env[62558]: DEBUG oslo_concurrency.lockutils [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] Acquiring lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.989115] env[62558]: DEBUG oslo_concurrency.lockutils [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] Acquired lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.989292] env[62558]: DEBUG nova.network.neutron [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Refreshing network info cache for port 2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.003792] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.003792] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521c1c5b-eeb5-f69f-d7f5-9c3d9d8e9ebb" [ 806.003792] env[62558]: _type = "HttpNfcLease" [ 806.003792] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 806.003792] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 806.003792] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521c1c5b-eeb5-f69f-d7f5-9c3d9d8e9ebb" [ 806.003792] env[62558]: _type = "HttpNfcLease" [ 806.003792] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 806.004356] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a793d7-3ec9-427c-a936-2ca345c3d9c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.016011] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 806.016011] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 806.074423] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.074690] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.074938] env[62558]: INFO nova.compute.manager [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Rebooting instance [ 806.126813] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d42edf9e-5924-4c95-b6e2-2476def21f1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.143376] env[62558]: DEBUG nova.compute.manager [req-58d880c4-aadd-4377-863f-3a7e935fdc38 req-9e97330b-2961-4c6a-a713-104c8234dd65 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Received event network-vif-deleted-c8245937-784e-4e48-8246-3323fab7a3dc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.143575] env[62558]: INFO nova.compute.manager [req-58d880c4-aadd-4377-863f-3a7e935fdc38 req-9e97330b-2961-4c6a-a713-104c8234dd65 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Neutron deleted interface c8245937-784e-4e48-8246-3323fab7a3dc; detaching it from the instance and deleting it from the info cache [ 806.143743] env[62558]: DEBUG nova.network.neutron [req-58d880c4-aadd-4377-863f-3a7e935fdc38 req-9e97330b-2961-4c6a-a713-104c8234dd65 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.203962] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266644, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.289765] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdfed20-dd9b-46da-be7a-491f61f52c8b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.297378] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a3fb3c-a759-463d-b5e8-6e0dd5b4e997 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.333833] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c01705-bd68-4fd6-ad40-96e35eeaedf6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.346513] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e74575-6e09-4166-85de-1a186e44d211 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.353146] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.363204] env[62558]: DEBUG nova.compute.provider_tree [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.441547] env[62558]: DEBUG oslo_vmware.api [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266640, 'name': PowerOnVM_Task, 'duration_secs': 0.529362} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.441998] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.442409] env[62558]: INFO nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Took 8.49 seconds to spawn the instance on the hypervisor. [ 806.442772] env[62558]: DEBUG nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.444013] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8e533f-8194-4258-b57a-1a6c76e05168 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.474493] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101471} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.474706] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.475542] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a924d52d-4464-471a-bc48-544ca79e7595 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.505712] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] a1d242e6-1561-4bd4-8e39-281ab6346661/a1d242e6-1561-4bd4-8e39-281ab6346661.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.509136] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28c128ec-6628-4f44-b120-cde0a1987649 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.532907] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 806.532907] env[62558]: value = "task-1266646" [ 806.532907] env[62558]: _type = "Task" [ 806.532907] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.542481] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.604193] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.604522] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquired lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.605197] env[62558]: DEBUG nova.network.neutron [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.615298] env[62558]: DEBUG nova.network.neutron [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.653318] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8265a07e-b810-4e62-8ed6-8d79d25f3d9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.663969] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36134611-fa92-49b4-9a4f-1d1497c88011 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.706068] env[62558]: DEBUG nova.compute.manager [req-58d880c4-aadd-4377-863f-3a7e935fdc38 req-9e97330b-2961-4c6a-a713-104c8234dd65 service nova] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Detach interface failed, port_id=c8245937-784e-4e48-8246-3323fab7a3dc, reason: Instance 3c8c2d25-981e-49de-b5c0-e0d12dd91378 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 806.715508] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266644, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.755131] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Successfully updated port: e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.823339] env[62558]: DEBUG nova.network.neutron [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Updated VIF entry in instance network info cache for port 2f4de907-3732-4ed6-a741-08fa0ce7db8f. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.823937] env[62558]: DEBUG nova.network.neutron [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Updating instance_info_cache with network_info: [{"id": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "address": "fa:16:3e:47:82:5f", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4de907-37", "ovs_interfaceid": "2f4de907-3732-4ed6-a741-08fa0ce7db8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.872494] env[62558]: DEBUG nova.scheduler.client.report [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.965217] env[62558]: INFO nova.compute.manager [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Took 39.53 seconds to build instance. [ 807.043916] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266646, 'name': ReconfigVM_Task, 'duration_secs': 0.502652} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.044427] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfigured VM instance instance-0000003c to attach disk [datastore1] a1d242e6-1561-4bd4-8e39-281ab6346661/a1d242e6-1561-4bd4-8e39-281ab6346661.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.045130] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30c73dbc-93ad-4e3d-9288-c497f42898b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.053491] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 807.053491] env[62558]: value = "task-1266647" [ 807.053491] env[62558]: _type = "Task" [ 807.053491] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.062506] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266647, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.119110] env[62558]: INFO nova.compute.manager [-] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Took 1.30 seconds to deallocate network for instance. [ 807.187744] env[62558]: INFO nova.compute.manager [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Rescuing [ 807.188518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.188518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.188518] env[62558]: DEBUG nova.network.neutron [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.214933] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266644, 'name': CreateVM_Task, 'duration_secs': 1.467191} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.215775] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.220019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.220019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.220019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.220019] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-372069ad-25cd-4b55-ae13-65fcc1b01b78 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.226099] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 807.226099] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f53c4f-14fe-4e8e-cea8-c3eef7f9b9c4" [ 807.226099] env[62558]: _type = "Task" [ 807.226099] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.237792] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f53c4f-14fe-4e8e-cea8-c3eef7f9b9c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.256991] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.257173] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.257330] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.327204] env[62558]: DEBUG oslo_concurrency.lockutils [req-03e892cb-8f4e-4c94-8815-c82edb909af5 req-b8cdb6e1-28bc-4235-8199-4cc620539ace service nova] Releasing lock "refresh_cache-ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.379634] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.722s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.379634] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.383546] env[62558]: DEBUG nova.network.neutron [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updating instance_info_cache with network_info: [{"id": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "address": "fa:16:3e:a9:64:2f", "network": {"id": "ad9fdca6-db0a-4237-8535-88789de4bc00", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-567677333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b336ad28239649b7949dd21ad2df85bd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4867bc9b-01", "ovs_interfaceid": "4867bc9b-0198-4b9c-ab0c-a41743956ee8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.384981] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.020s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.385406] env[62558]: DEBUG nova.objects.instance [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'resources' on Instance uuid b0038711-5d12-4909-b331-72acb5ed0d24 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.467593] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6647f882-579d-4788-af6d-27552ee60487 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.326s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.564509] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266647, 'name': Rename_Task, 'duration_secs': 0.26647} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.565126] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.565435] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df60c897-c64f-45f7-aab0-740800066a95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.572549] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 807.572549] env[62558]: value = "task-1266648" [ 807.572549] env[62558]: _type = "Task" [ 807.572549] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.580653] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266648, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.625771] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.738638] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f53c4f-14fe-4e8e-cea8-c3eef7f9b9c4, 'name': SearchDatastore_Task, 'duration_secs': 0.019342} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.739020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.739277] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.739521] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.739688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.740036] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.740344] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bccce265-bdc3-4b2a-b6e0-86f4ba9ec9aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.749542] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.749749] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.750492] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a4d93d0-190c-46f2-94da-5535a95c83d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.755816] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 807.755816] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fa2ae3-1e39-f3c8-2953-273bf271d815" [ 807.755816] env[62558]: _type = "Task" [ 807.755816] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.763877] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fa2ae3-1e39-f3c8-2953-273bf271d815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.815900] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.889199] env[62558]: DEBUG nova.compute.utils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.895025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Releasing lock "refresh_cache-a8fef137-a651-4724-8ad1-a3e6ddcb5b03" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.895025] env[62558]: DEBUG nova.compute.manager [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.900087] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.900792] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 807.903379] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334e41f2-f915-4631-96f1-b272ebe2ad8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.970717] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.981574] env[62558]: DEBUG nova.policy [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e6e2f1bae4d4bb68070770dbbbe5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78b69738b90142a4943069006e349445', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.986455] env[62558]: DEBUG nova.network.neutron [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.040269] env[62558]: DEBUG nova.compute.manager [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Received event network-vif-plugged-e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.040560] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Acquiring lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.041090] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.041090] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.041248] env[62558]: DEBUG nova.compute.manager [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] No waiting events found dispatching network-vif-plugged-e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.041414] env[62558]: WARNING nova.compute.manager [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Received unexpected event network-vif-plugged-e384e059-74fe-48a6-a03e-e6262dee7821 for instance with vm_state building and task_state spawning. [ 808.041599] env[62558]: DEBUG nova.compute.manager [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Received event network-changed-e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.041776] env[62558]: DEBUG nova.compute.manager [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Refreshing instance network info cache due to event network-changed-e384e059-74fe-48a6-a03e-e6262dee7821. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.041978] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Acquiring lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.046814] env[62558]: DEBUG nova.network.neutron [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Updating instance_info_cache with network_info: [{"id": "e384e059-74fe-48a6-a03e-e6262dee7821", "address": "fa:16:3e:f7:26:a7", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape384e059-74", "ovs_interfaceid": "e384e059-74fe-48a6-a03e-e6262dee7821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.087064] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266648, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.201637] env[62558]: INFO nova.compute.manager [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Rebuilding instance [ 808.254720] env[62558]: DEBUG nova.compute.manager [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.255663] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eeb38bd-d002-4095-90bf-40ce6e1a431b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.270937] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fa2ae3-1e39-f3c8-2953-273bf271d815, 'name': SearchDatastore_Task, 'duration_secs': 0.015112} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.277156] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e804d53e-2813-44b0-89c5-88d737dc1a84 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.284116] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 808.284116] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f70c49-3f3e-68a3-6c89-820bcc0c0e28" [ 808.284116] env[62558]: _type = "Task" [ 808.284116] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.296439] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f70c49-3f3e-68a3-6c89-820bcc0c0e28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.310580] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Successfully created port: 6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.340642] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff65023-1ec1-4d6b-9cc5-fae2e062cd87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.350747] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919e5fb0-0d35-457e-971a-9de5e8f672ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.386603] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0240567a-1b16-46aa-8ee6-d7198c7bda41 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.394521] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665bb9ff-fe52-4058-9136-9a23c4b76a38 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.399327] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.410157] env[62558]: DEBUG nova.compute.provider_tree [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.489207] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.497719] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.553682] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.553991] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Instance network_info: |[{"id": "e384e059-74fe-48a6-a03e-e6262dee7821", "address": "fa:16:3e:f7:26:a7", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape384e059-74", "ovs_interfaceid": "e384e059-74fe-48a6-a03e-e6262dee7821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.554380] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Acquired lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.554571] env[62558]: DEBUG nova.network.neutron [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Refreshing network info cache for port e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.555831] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:26:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e384e059-74fe-48a6-a03e-e6262dee7821', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.564245] env[62558]: DEBUG oslo.service.loopingcall [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.567714] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.568232] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65570eb3-518a-4494-ba33-20e59d264afe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.592653] env[62558]: DEBUG oslo_vmware.api [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266648, 'name': PowerOnVM_Task, 'duration_secs': 0.683074} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.594083] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.594330] env[62558]: INFO nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Took 8.03 seconds to spawn the instance on the hypervisor. [ 808.594507] env[62558]: DEBUG nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.594795] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.594795] env[62558]: value = "task-1266649" [ 808.594795] env[62558]: _type = "Task" [ 808.594795] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.595540] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eae3926-939a-410d-a448-7b867152dd20 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.606448] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266649, 'name': CreateVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.782059] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 808.782433] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a667461-3eea-4e4c-8df2-bc68ce24c619 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.792068] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 808.792068] env[62558]: value = "task-1266650" [ 808.792068] env[62558]: _type = "Task" [ 808.792068] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.800922] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f70c49-3f3e-68a3-6c89-820bcc0c0e28, 'name': SearchDatastore_Task, 'duration_secs': 0.016532} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.801651] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.801945] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55/ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.802243] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1939ebca-8b60-4767-9ef7-61117a894753 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.808944] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.813574] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 808.813574] env[62558]: value = "task-1266651" [ 808.813574] env[62558]: _type = "Task" [ 808.813574] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.821719] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266651, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.886674] env[62558]: DEBUG nova.network.neutron [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Updated VIF entry in instance network info cache for port e384e059-74fe-48a6-a03e-e6262dee7821. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 808.887842] env[62558]: DEBUG nova.network.neutron [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Updating instance_info_cache with network_info: [{"id": "e384e059-74fe-48a6-a03e-e6262dee7821", "address": "fa:16:3e:f7:26:a7", "network": {"id": "be055322-c2ca-45bd-a90f-227e4f6c5d99", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-901481067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9fabbd2127c2445aa96cf0d513541d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape384e059-74", "ovs_interfaceid": "e384e059-74fe-48a6-a03e-e6262dee7821", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.918030] env[62558]: DEBUG nova.scheduler.client.report [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.930512] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3ac2f2-c88e-440a-b72d-14a750d073ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.948084] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Doing hard reboot of VM {{(pid=62558) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 808.948564] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-e725dfc8-0769-42e0-a025-31e6d42632cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.955355] env[62558]: DEBUG oslo_vmware.api [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 808.955355] env[62558]: value = "task-1266652" [ 808.955355] env[62558]: _type = "Task" [ 808.955355] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.966081] env[62558]: DEBUG oslo_vmware.api [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266652, 'name': ResetVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.019169] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.019669] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba727c05-148b-4fd9-96c1-0ddffb946aa2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.030052] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 809.030052] env[62558]: value = "task-1266653" [ 809.030052] env[62558]: _type = "Task" [ 809.030052] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.038647] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.109034] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266649, 'name': CreateVM_Task, 'duration_secs': 0.366001} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.109034] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.110427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.110427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.110427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.110718] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fe4f30e-71b8-4eef-9347-518a5651aeb9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.124539] env[62558]: INFO nova.compute.manager [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Took 40.44 seconds to build instance. [ 809.126211] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 809.126211] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cdd273-7bc9-625a-089d-340fe8f25af5" [ 809.126211] env[62558]: _type = "Task" [ 809.126211] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.139488] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cdd273-7bc9-625a-089d-340fe8f25af5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.303826] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266650, 'name': PowerOffVM_Task, 'duration_secs': 0.190058} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.304128] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.304389] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.305221] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53ab470-b537-4b3c-9342-104213ef57f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.314416] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.317944] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89acb821-3cf7-46f6-a21b-b59f13b5f29e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.325330] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266651, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.392204] env[62558]: DEBUG oslo_concurrency.lockutils [req-5d333505-0f4b-4376-943e-543403eb97ae req-be881c68-ee4c-45ff-9f56-f5231b86f98d service nova] Releasing lock "refresh_cache-ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.423144] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.426721] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.429837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.108s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.430406] env[62558]: INFO nova.compute.claims [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.461022] env[62558]: INFO nova.scheduler.client.report [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted allocations for instance b0038711-5d12-4909-b331-72acb5ed0d24 [ 809.468317] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.471219] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.471219] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.471219] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.471219] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.471219] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.471465] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.471465] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.471465] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.471465] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.471465] env[62558]: DEBUG nova.virt.hardware [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.471598] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dd21a5-4308-4622-8729-7a02a83742e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.480484] env[62558]: DEBUG oslo_vmware.api [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266652, 'name': ResetVM_Task, 'duration_secs': 0.109919} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.482897] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Did hard reboot of VM {{(pid=62558) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 809.483121] env[62558]: DEBUG nova.compute.manager [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.483941] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d3c1b5-fd45-43f5-a650-7d0a30cc4f5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.488014] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618ad285-5ee3-48a7-b4c9-56decb09b021 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.542675] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266653, 'name': PowerOffVM_Task, 'duration_secs': 0.445916} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.542950] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.543761] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32954b0-0f78-4edf-b475-7cede2e2b0c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.564484] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9397a86-b54d-4efb-94d3-3cdd2bfd01f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.593048] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.593048] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86b780b9-83ef-428a-ac87-893d92d0dfb2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.600230] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 809.600230] env[62558]: value = "task-1266655" [ 809.600230] env[62558]: _type = "Task" [ 809.600230] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.607851] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266655, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.628029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a663a5d8-5843-4b81-abfd-6b750e5c920e tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.988s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.638622] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cdd273-7bc9-625a-089d-340fe8f25af5, 'name': SearchDatastore_Task, 'duration_secs': 0.024332} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.639525] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.639771] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.640029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.640599] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.640599] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.641559] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a22e813-ea9f-417e-bd99-d06a76dee4c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.649945] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.650166] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.650923] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc437dcf-1d05-4f2b-9bfe-8b81c54cd826 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.656734] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 809.656734] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f2dd9c-8484-ab57-585d-6c5431b88c76" [ 809.656734] env[62558]: _type = "Task" [ 809.656734] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.664280] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f2dd9c-8484-ab57-585d-6c5431b88c76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.823819] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266651, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597283} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.824305] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55/ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.824380] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.824617] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f8eb23b-f520-4432-8a14-d8e9637bd472 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.832688] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 809.832688] env[62558]: value = "task-1266656" [ 809.832688] env[62558]: _type = "Task" [ 809.832688] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.839942] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.924670] env[62558]: DEBUG nova.compute.manager [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received event network-vif-plugged-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.925436] env[62558]: DEBUG oslo_concurrency.lockutils [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] Acquiring lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.925696] env[62558]: DEBUG oslo_concurrency.lockutils [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.925892] env[62558]: DEBUG oslo_concurrency.lockutils [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.926083] env[62558]: DEBUG nova.compute.manager [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] No waiting events found dispatching network-vif-plugged-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.926266] env[62558]: WARNING nova.compute.manager [req-c7dadc3f-ad7d-4076-bea2-cf6c126c6d02 req-99a2b234-a685-4ca4-b221-70e5386f4508 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received unexpected event network-vif-plugged-6fcb24d4-f48e-4628-995d-10f53f01154b for instance with vm_state building and task_state spawning. [ 809.978903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-07802eaf-76af-4e6d-a1dc-9ab2c0e681b1 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "b0038711-5d12-4909-b331-72acb5ed0d24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.003302] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46bbf025-d971-4325-ab92-7398d2223ca7 tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.928s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.112376] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 810.112613] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.112868] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.113023] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.113268] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.113476] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bac753e9-6fea-4146-80e9-066fdeff4f29 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.124809] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.124809] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.124809] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd01e681-5113-48e0-bea9-e490a19a971f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.129922] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 810.129922] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52576db3-4167-f7e1-1a6f-499447d91f42" [ 810.129922] env[62558]: _type = "Task" [ 810.129922] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.133278] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.140354] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52576db3-4167-f7e1-1a6f-499447d91f42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.151601] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.151828] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.152015] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleting the datastore file [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.152294] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a3cd022-13cc-4721-87dd-4aa5375cb564 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.158472] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 810.158472] env[62558]: value = "task-1266657" [ 810.158472] env[62558]: _type = "Task" [ 810.158472] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.171230] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f2dd9c-8484-ab57-585d-6c5431b88c76, 'name': SearchDatastore_Task, 'duration_secs': 0.030284} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.175747] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266657, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.176246] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a74e29a1-d5a9-4928-b8fa-ecaddb7f7b71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.182314] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 810.182314] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f41101-a842-1ed9-b848-8f0715493635" [ 810.182314] env[62558]: _type = "Task" [ 810.182314] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.191953] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f41101-a842-1ed9-b848-8f0715493635, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.298955] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.299492] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.299861] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.300213] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.300514] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.303775] env[62558]: INFO nova.compute.manager [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Terminating instance [ 810.306672] env[62558]: DEBUG nova.compute.manager [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 810.307012] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.308469] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8848a32b-6579-4790-a83d-2d964e909930 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.318198] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.318759] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d222af81-0c08-4a03-a7d3-303510db8ddf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.325039] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 810.325039] env[62558]: value = "task-1266658" [ 810.325039] env[62558]: _type = "Task" [ 810.325039] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.334461] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.335430] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Successfully updated port: 6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.345288] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155704} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.345556] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.346421] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3cec85-b4b8-4e2b-b3fc-8c19b297b91a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.371606] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55/ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.372366] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee0055aa-53a0-49da-b89d-0e41cc76b3c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.394077] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 810.394077] env[62558]: value = "task-1266659" [ 810.394077] env[62558]: _type = "Task" [ 810.394077] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.404115] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266659, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.568625] env[62558]: DEBUG nova.compute.manager [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.568860] env[62558]: DEBUG nova.compute.manager [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing instance network info cache due to event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.569125] env[62558]: DEBUG oslo_concurrency.lockutils [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] Acquiring lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.569284] env[62558]: DEBUG oslo_concurrency.lockutils [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] Acquired lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.569470] env[62558]: DEBUG nova.network.neutron [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing network info cache for port 6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 810.645936] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52576db3-4167-f7e1-1a6f-499447d91f42, 'name': SearchDatastore_Task, 'duration_secs': 0.024933} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.649241] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe67abfb-3222-4303-927b-f3944d505abb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.654675] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 810.654675] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523251e5-bc37-dad8-fd98-66ce846348cd" [ 810.654675] env[62558]: _type = "Task" [ 810.654675] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.663578] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.670874] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523251e5-bc37-dad8-fd98-66ce846348cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.675696] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266657, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348857} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.678850] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.679109] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.679369] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.693025] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f41101-a842-1ed9-b848-8f0715493635, 'name': SearchDatastore_Task, 'duration_secs': 0.023153} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.693417] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.693772] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ebd84b3f-0c5d-40e3-aa70-f8b3054b7109/ebd84b3f-0c5d-40e3-aa70-f8b3054b7109.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.694105] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44b18b61-5c37-4123-8a42-e05c13a92f1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.704924] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 810.704924] env[62558]: value = "task-1266660" [ 810.704924] env[62558]: _type = "Task" [ 810.704924] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.713705] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.838438] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266658, 'name': PowerOffVM_Task, 'duration_secs': 0.236976} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.842109] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.842359] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.842985] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.843451] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26be3ed4-22f7-419b-abc5-fa8fce9ad2a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.861018] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9e9972-814e-40d0-bc07-2df38b9e3a9b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.872836] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62bc68b-d3ed-4640-9d9c-34eb1cfafe9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.915021] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a332553-7c5f-4509-9184-7d09f95c0195 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.932949] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266659, 'name': ReconfigVM_Task, 'duration_secs': 0.476387} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.934812] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3205d7a-787f-4e56-8f7e-95c278d3d1c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.941360] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Reconfigured VM instance instance-0000003d to attach disk [datastore1] ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55/ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.943381] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffb83ae7-c5d2-4748-816e-822515dee072 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.949150] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.949150] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.949150] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleting the datastore file [datastore2] 6e01c855-ef25-45c9-a73e-5656e8e05cd4 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.949150] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c32981f1-ac78-4dae-89a7-4626f8df5645 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.967111] env[62558]: DEBUG nova.compute.provider_tree [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.972080] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 810.972080] env[62558]: value = "task-1266662" [ 810.972080] env[62558]: _type = "Task" [ 810.972080] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.978629] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 810.978629] env[62558]: value = "task-1266663" [ 810.978629] env[62558]: _type = "Task" [ 810.978629] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.992246] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266662, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.000714] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.123073] env[62558]: DEBUG nova.network.neutron [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.189616] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523251e5-bc37-dad8-fd98-66ce846348cd, 'name': SearchDatastore_Task, 'duration_secs': 0.014003} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.192965] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.193506] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. {{(pid=62558) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 811.194259] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e236a5ef-99f8-4d27-976d-4c3b8c40426b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.202145] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 811.202145] env[62558]: value = "task-1266664" [ 811.202145] env[62558]: _type = "Task" [ 811.202145] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.228033] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495353} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.228316] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.228632] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ebd84b3f-0c5d-40e3-aa70-f8b3054b7109/ebd84b3f-0c5d-40e3-aa70-f8b3054b7109.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.228875] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.229203] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5962f913-4be7-462d-bd38-e416e52757cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.238919] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 811.238919] env[62558]: value = "task-1266665" [ 811.238919] env[62558]: _type = "Task" [ 811.238919] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.248796] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.251987] env[62558]: DEBUG nova.compute.manager [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Received event network-changed-bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.252403] env[62558]: DEBUG nova.compute.manager [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Refreshing instance network info cache due to event network-changed-bdbc998f-cabe-438b-b3ce-6477b78a13e7. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.252481] env[62558]: DEBUG oslo_concurrency.lockutils [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] Acquiring lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.252654] env[62558]: DEBUG oslo_concurrency.lockutils [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] Acquired lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.252818] env[62558]: DEBUG nova.network.neutron [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Refreshing network info cache for port bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.255167] env[62558]: DEBUG nova.network.neutron [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.483914] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266662, 'name': Rename_Task, 'duration_secs': 0.217569} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.487214] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.487528] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-829d695d-8f61-4842-bf1a-67ca50edf0fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.498338] env[62558]: DEBUG oslo_vmware.api [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339044} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.499830] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.500051] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.500259] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.500439] env[62558]: INFO nova.compute.manager [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Took 1.19 seconds to destroy the instance on the hypervisor. [ 811.500732] env[62558]: DEBUG oslo.service.loopingcall [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.501015] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 811.501015] env[62558]: value = "task-1266666" [ 811.501015] env[62558]: _type = "Task" [ 811.501015] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.501222] env[62558]: DEBUG nova.compute.manager [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.501315] env[62558]: DEBUG nova.network.neutron [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.507670] env[62558]: DEBUG nova.scheduler.client.report [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 811.508072] env[62558]: DEBUG nova.compute.provider_tree [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 85 to 86 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 811.508363] env[62558]: DEBUG nova.compute.provider_tree [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.515600] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266666, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.723097] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266664, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.727935] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.727935] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.728171] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.728507] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.728780] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.728910] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.729344] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.729581] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.729793] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.730042] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.730236] env[62558]: DEBUG nova.virt.hardware [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.731486] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b50ff9-537f-4f3b-8743-c6bfd388fa89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.746162] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73690fb4-1c60-4363-8d3e-3f5827b962e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.768911] env[62558]: DEBUG oslo_concurrency.lockutils [req-457f8810-a948-4891-bd0f-b7d3a2d5032f req-e1d99d9b-ffc8-4ae5-8a5c-ca1234e96441 service nova] Releasing lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.772265] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:2e:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f256cfee-512d-4192-9aca-6750fdb1cd4c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc1b2699-3507-45b0-afe0-572fd6966fdf', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.782273] env[62558]: DEBUG oslo.service.loopingcall [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.785128] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.785366] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.788262] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.788262] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071672} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.788262] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cfb9bc83-2535-4300-86c1-1011ee5d642e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.803086] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.805881] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa4de87-7434-4de6-ad20-67935c0639b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.831896] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] ebd84b3f-0c5d-40e3-aa70-f8b3054b7109/ebd84b3f-0c5d-40e3-aa70-f8b3054b7109.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.835360] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-969ac025-88b6-4b1e-843f-641585f5212b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.849272] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.849272] env[62558]: value = "task-1266667" [ 811.849272] env[62558]: _type = "Task" [ 811.849272] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.849999] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.850255] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.850461] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.850676] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.850863] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.853385] env[62558]: INFO nova.compute.manager [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Terminating instance [ 811.858330] env[62558]: DEBUG nova.compute.manager [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 811.858521] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 811.860270] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e951cc-2ab9-4d59-aaa5-dfb384be4841 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.863478] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 811.863478] env[62558]: value = "task-1266668" [ 811.863478] env[62558]: _type = "Task" [ 811.863478] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.870650] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266667, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.872805] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 811.875802] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1bfb94f-e005-422d-acf2-7d01d56f2ff1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.877392] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266668, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.882937] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 811.882937] env[62558]: value = "task-1266669" [ 811.882937] env[62558]: _type = "Task" [ 811.882937] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.892193] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.016851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.017393] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.020244] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.020666] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.530s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.022451] env[62558]: INFO nova.compute.claims [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.215198] env[62558]: DEBUG nova.network.neutron [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updated VIF entry in instance network info cache for port bdbc998f-cabe-438b-b3ce-6477b78a13e7. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.215719] env[62558]: DEBUG nova.network.neutron [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.230680] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667091} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.231033] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. [ 812.231902] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e608513b-ee8b-4044-a03e-6fd13a2ecc84 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.259528] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.260289] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd9a8b04-b20b-4f3b-8fa1-1d8f741b6645 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.280778] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 812.280778] env[62558]: value = "task-1266670" [ 812.280778] env[62558]: _type = "Task" [ 812.280778] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.290806] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266670, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.309989] env[62558]: DEBUG nova.network.neutron [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.345224] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.362631] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266667, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.376640] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266668, 'name': ReconfigVM_Task, 'duration_secs': 0.380204} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.377083] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Reconfigured VM instance instance-0000003e to attach disk [datastore2] ebd84b3f-0c5d-40e3-aa70-f8b3054b7109/ebd84b3f-0c5d-40e3-aa70-f8b3054b7109.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.377859] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2ae2b6d-cb35-4c56-9166-a5423604dfc3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.384607] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 812.384607] env[62558]: value = "task-1266671" [ 812.384607] env[62558]: _type = "Task" [ 812.384607] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.396391] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266669, 'name': PowerOffVM_Task, 'duration_secs': 0.304035} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.399785] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.399998] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 812.400295] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266671, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.400535] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f28252a-4730-4239-9329-3dd51f805dda {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.477324] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.477665] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.477935] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Deleting the datastore file [datastore2] a8fef137-a651-4724-8ad1-a3e6ddcb5b03 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.478407] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92ec718c-3afc-427d-bd8b-4d4a7c0ad0dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.485295] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for the task: (returnval){ [ 812.485295] env[62558]: value = "task-1266673" [ 812.485295] env[62558]: _type = "Task" [ 812.485295] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.493801] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.517323] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.527952] env[62558]: DEBUG nova.compute.utils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.531461] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.531702] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.575729] env[62558]: DEBUG nova.network.neutron [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [{"id": "6fcb24d4-f48e-4628-995d-10f53f01154b", "address": "fa:16:3e:09:d7:9e", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb24d4-f4", "ovs_interfaceid": "6fcb24d4-f48e-4628-995d-10f53f01154b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.578170] env[62558]: DEBUG nova.policy [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3c4474feb3a4f5ea478ad5e30d9ea49', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '480afedc5a8845e8920815407f4485dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.599199] env[62558]: DEBUG nova.compute.manager [req-c1de2472-4fcf-4143-965f-a7492bb22900 req-0423dd36-759a-43d7-88b3-08610d77fcfe service nova] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Received event network-vif-deleted-9245beb2-0701-4744-b916-d8291b7b8db0 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.718754] env[62558]: DEBUG oslo_concurrency.lockutils [req-d13786a5-e225-4e58-b5ee-d87c4f646643 req-5d864472-4a7e-4b9d-b689-b03ecb35b1dd service nova] Releasing lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.792209] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266670, 'name': ReconfigVM_Task, 'duration_secs': 0.404775} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.792522] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.793359] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e558488-dd97-4525-93b7-725ba3ffcfc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.813814] env[62558]: INFO nova.compute.manager [-] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Took 1.31 seconds to deallocate network for instance. [ 812.821022] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-380644f8-27b7-4bbe-add0-0c40b4e8710a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.842477] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 812.842477] env[62558]: value = "task-1266674" [ 812.842477] env[62558]: _type = "Task" [ 812.842477] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.850808] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266674, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.865584] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266667, 'name': CreateVM_Task, 'duration_secs': 0.711336} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.865771] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.866516] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.866687] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.867018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.867375] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df7d045a-e4cf-4f2f-b474-8d24154350f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.881158] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 812.881158] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3d824-9b30-2a02-7771-1e89b7d61765" [ 812.881158] env[62558]: _type = "Task" [ 812.881158] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.886779] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Successfully created port: 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.895401] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3d824-9b30-2a02-7771-1e89b7d61765, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.900707] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266671, 'name': Rename_Task, 'duration_secs': 0.151577} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.900970] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.901272] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d38b648-b8ac-4d9e-8c95-c86b0e1cb6bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.907984] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 812.907984] env[62558]: value = "task-1266675" [ 812.907984] env[62558]: _type = "Task" [ 812.907984] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.916340] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.996272] env[62558]: DEBUG oslo_vmware.api [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Task: {'id': task-1266673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182119} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.996540] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.996782] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 812.997016] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 812.997261] env[62558]: INFO nova.compute.manager [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Took 1.14 seconds to destroy the instance on the hypervisor. [ 812.997562] env[62558]: DEBUG oslo.service.loopingcall [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.997860] env[62558]: DEBUG nova.compute.manager [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.997979] env[62558]: DEBUG nova.network.neutron [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.015395] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266666, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.032802] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.081023] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.081353] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance network_info: |[{"id": "6fcb24d4-f48e-4628-995d-10f53f01154b", "address": "fa:16:3e:09:d7:9e", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb24d4-f4", "ovs_interfaceid": "6fcb24d4-f48e-4628-995d-10f53f01154b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.082048] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:d7:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fcb24d4-f48e-4628-995d-10f53f01154b', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.093266] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating folder: Project (78b69738b90142a4943069006e349445). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.097832] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b308dec-810a-4770-9ffb-a873853f00a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.108299] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created folder: Project (78b69738b90142a4943069006e349445) in parent group-v272451. [ 813.108627] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating folder: Instances. Parent ref: group-v272503. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.108918] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0aeb2d1e-e34e-4f9a-88ff-6ab055100973 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.119302] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created folder: Instances in parent group-v272503. [ 813.119562] env[62558]: DEBUG oslo.service.loopingcall [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.119758] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.119977] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3756579a-5a49-4cbe-834d-64b2b9827d31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.144251] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.144251] env[62558]: value = "task-1266678" [ 813.144251] env[62558]: _type = "Task" [ 813.144251] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.158690] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266678, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.340596] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.356438] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266674, 'name': ReconfigVM_Task, 'duration_secs': 0.19692} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.356724] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.356997] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23e63088-1688-48cc-9da3-f3d1e5abd290 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.364923] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 813.364923] env[62558]: value = "task-1266679" [ 813.364923] env[62558]: _type = "Task" [ 813.364923] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.376924] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266679, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.395381] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3d824-9b30-2a02-7771-1e89b7d61765, 'name': SearchDatastore_Task, 'duration_secs': 0.011427} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.395849] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.396223] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.396578] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.396904] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.397769] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.397769] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8758464f-5525-414a-9742-9f2ee50c235c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.415566] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.419020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.420638] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3602a04-0423-4851-bdf8-850a38ea74b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.432181] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266675, 'name': PowerOnVM_Task, 'duration_secs': 0.475878} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.432181] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.432339] env[62558]: INFO nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Took 7.74 seconds to spawn the instance on the hypervisor. [ 813.433455] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.433739] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0d4ec2-2f77-483f-85e7-79b8982a0f90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.439366] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 813.439366] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521b957a-106d-fdb6-2688-6d32eeccbebf" [ 813.439366] env[62558]: _type = "Task" [ 813.439366] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.457912] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521b957a-106d-fdb6-2688-6d32eeccbebf, 'name': SearchDatastore_Task, 'duration_secs': 0.011885} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.458740] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19a3070f-d2b5-4ce6-acde-6caa776a32d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.464411] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 813.464411] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52953b3f-48a1-61cf-97e0-01f28daf064e" [ 813.464411] env[62558]: _type = "Task" [ 813.464411] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.477027] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52953b3f-48a1-61cf-97e0-01f28daf064e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.515509] env[62558]: DEBUG oslo_vmware.api [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266666, 'name': PowerOnVM_Task, 'duration_secs': 1.543703} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.517593] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.517930] env[62558]: INFO nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Took 10.37 seconds to spawn the instance on the hypervisor. [ 813.518230] env[62558]: DEBUG nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.519480] env[62558]: DEBUG nova.compute.manager [req-df28ecb5-97d9-49c4-85e3-6ca4e5fc86e9 req-55bbdd61-70a0-4445-991b-a58ff6f9fff9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Received event network-vif-deleted-4867bc9b-0198-4b9c-ab0c-a41743956ee8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.519685] env[62558]: INFO nova.compute.manager [req-df28ecb5-97d9-49c4-85e3-6ca4e5fc86e9 req-55bbdd61-70a0-4445-991b-a58ff6f9fff9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Neutron deleted interface 4867bc9b-0198-4b9c-ab0c-a41743956ee8; detaching it from the instance and deleting it from the info cache [ 813.519862] env[62558]: DEBUG nova.network.neutron [req-df28ecb5-97d9-49c4-85e3-6ca4e5fc86e9 req-55bbdd61-70a0-4445-991b-a58ff6f9fff9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.522809] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0561f62c-dcd2-4523-8f48-bb75818dfd2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.526444] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5136f1-8a3c-4103-8be5-7e05c4c013d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.536591] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993e597d-1674-4d6d-b383-66e6964f1e9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.576995] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202312e2-ecd5-4466-ac81-bf9b30775559 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.586195] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0160f0c-b9c5-4735-a243-f5d6f694ff4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.602463] env[62558]: DEBUG nova.compute.provider_tree [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.653767] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266678, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.874714] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266679, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.933041] env[62558]: DEBUG nova.network.neutron [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.963353] env[62558]: INFO nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Took 43.42 seconds to build instance. [ 813.975240] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52953b3f-48a1-61cf-97e0-01f28daf064e, 'name': SearchDatastore_Task, 'duration_secs': 0.01851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.975540] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.975802] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.976094] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6962b7ff-171a-4f54-ba6d-5a4b7c62b185 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.983894] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 813.983894] env[62558]: value = "task-1266680" [ 813.983894] env[62558]: _type = "Task" [ 813.983894] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.992894] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.030839] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6558497d-cf27-45bb-aa95-c845146fd222 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.040237] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d16e1d-cfdf-41cf-90d8-2a3698ec3d28 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.058023] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.060203] env[62558]: INFO nova.compute.manager [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Took 43.84 seconds to build instance. [ 814.078173] env[62558]: DEBUG nova.compute.manager [req-df28ecb5-97d9-49c4-85e3-6ca4e5fc86e9 req-55bbdd61-70a0-4445-991b-a58ff6f9fff9 service nova] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Detach interface failed, port_id=4867bc9b-0198-4b9c-ab0c-a41743956ee8, reason: Instance a8fef137-a651-4724-8ad1-a3e6ddcb5b03 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 814.138729] env[62558]: DEBUG nova.scheduler.client.report [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 814.139075] env[62558]: DEBUG nova.compute.provider_tree [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 86 to 87 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 814.139266] env[62558]: DEBUG nova.compute.provider_tree [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.157503] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266678, 'name': CreateVM_Task, 'duration_secs': 0.537928} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.157686] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.158706] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.158706] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.159073] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.159349] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1be429c-33f9-4205-8d7d-22150a23c1c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.164200] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.164453] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.164634] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.164842] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.164997] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.165158] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.165390] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.165564] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.165741] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.165923] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.167294] env[62558]: DEBUG nova.virt.hardware [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.167294] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1311137f-4a33-4735-bd23-45383f2335c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.172974] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 814.172974] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5287376d-202b-efc1-abde-600e388e02b3" [ 814.172974] env[62558]: _type = "Task" [ 814.172974] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.181345] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f7bb7c-3628-4e34-9a68-27e238f7e1e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.189263] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5287376d-202b-efc1-abde-600e388e02b3, 'name': SearchDatastore_Task, 'duration_secs': 0.016238} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.189965] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.190352] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.190745] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.190952] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.191241] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.200623] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b9c29c7-4b09-491b-84f4-cf045f466c26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.207591] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.207591] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 814.208504] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d24f742-fb05-482b-bf50-35a1c52db4e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.213276] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 814.213276] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528ee0f4-6a7e-e846-c6c1-0bbf938687b0" [ 814.213276] env[62558]: _type = "Task" [ 814.213276] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.221338] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528ee0f4-6a7e-e846-c6c1-0bbf938687b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.376779] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266679, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.435568] env[62558]: INFO nova.compute.manager [-] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Took 1.44 seconds to deallocate network for instance. [ 814.465447] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.769s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.497345] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266680, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.562981] env[62558]: DEBUG oslo_concurrency.lockutils [None req-425d31a9-8623-40f5-822f-74fe4f3e4e19 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.901s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.645486] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.646046] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.649280] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.553s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.650798] env[62558]: INFO nova.compute.claims [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.672591] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Successfully updated port: 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.694041] env[62558]: DEBUG nova.compute.manager [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.694152] env[62558]: DEBUG oslo_concurrency.lockutils [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.694465] env[62558]: DEBUG oslo_concurrency.lockutils [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.694729] env[62558]: DEBUG oslo_concurrency.lockutils [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.694949] env[62558]: DEBUG nova.compute.manager [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] No waiting events found dispatching network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.695178] env[62558]: WARNING nova.compute.manager [req-867a94a9-e167-4870-b2d4-bc7ddf449f7f req-fefe5bc9-cdfa-4cf6-a0e1-976436436a4b service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received unexpected event network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b for instance with vm_state building and task_state spawning. [ 814.724014] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528ee0f4-6a7e-e846-c6c1-0bbf938687b0, 'name': SearchDatastore_Task, 'duration_secs': 0.012657} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.724840] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07f46a5-b6af-419b-9feb-13cac5d4d809 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.730371] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 814.730371] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522f407c-2761-402c-bae8-7cd7516170cf" [ 814.730371] env[62558]: _type = "Task" [ 814.730371] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.739456] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522f407c-2761-402c-bae8-7cd7516170cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.877601] env[62558]: DEBUG oslo_vmware.api [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266679, 'name': PowerOnVM_Task, 'duration_secs': 1.105763} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.877911] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.880902] env[62558]: DEBUG nova.compute.manager [None req-9807ea24-ea72-4fe4-a5e8-7ec4a2ba9ecd tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.881838] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add06433-b77e-4935-a013-5b5071be2a22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.947408] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.968397] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.995895] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604163} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.996135] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.996354] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.996600] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44063078-9f0e-414a-89a8-6d7dd3462674 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.004372] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 815.004372] env[62558]: value = "task-1266681" [ 815.004372] env[62558]: _type = "Task" [ 815.004372] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.013290] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266681, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.065499] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.155791] env[62558]: DEBUG nova.compute.utils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.159507] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.159643] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.176562] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.176711] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.176853] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.213873] env[62558]: DEBUG nova.policy [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61cb5cd30d434f1eb594dd4521dab2d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e1d3e1dcd64e918c1e436e691e290b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.241314] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522f407c-2761-402c-bae8-7cd7516170cf, 'name': SearchDatastore_Task, 'duration_secs': 0.019767} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.241593] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.241885] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b/4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 815.242175] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6e0ad62-05c7-45aa-a75e-c0ba6844d963 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.248857] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 815.248857] env[62558]: value = "task-1266682" [ 815.248857] env[62558]: _type = "Task" [ 815.248857] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.257992] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.479736] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Successfully created port: 8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.485268] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.485480] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.485520] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.485712] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.486392] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.488163] env[62558]: INFO nova.compute.manager [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Terminating instance [ 815.489769] env[62558]: DEBUG nova.compute.manager [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.489976] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.491203] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4d39e2-e195-44ee-ba9a-b474189e7100 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.495047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.500490] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.500490] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed43b10a-e4aa-4d09-bdbd-d28164403c29 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.511972] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 815.511972] env[62558]: value = "task-1266683" [ 815.511972] env[62558]: _type = "Task" [ 815.511972] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.518552] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266681, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070181} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.519269] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.520103] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aa5902-0477-4368-be03-9c5e18c8a1f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.525486] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.546053] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.546450] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad85fbe1-bada-4956-a11d-bfca365a98d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.562973] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.563289] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.563498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.563681] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.563840] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.566339] env[62558]: INFO nova.compute.manager [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Terminating instance [ 815.568232] env[62558]: DEBUG nova.compute.manager [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.568422] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.571091] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c1f338-2fbc-4772-a342-39af49678fcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.577065] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 815.577065] env[62558]: value = "task-1266684" [ 815.577065] env[62558]: _type = "Task" [ 815.577065] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.583083] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.583732] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c17f3d54-4bd7-459b-acf6-b7da4b550277 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.588432] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.592459] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.593938] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 815.593938] env[62558]: value = "task-1266685" [ 815.593938] env[62558]: _type = "Task" [ 815.593938] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.603038] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.660659] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.749555] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.766491] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266682, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.898722] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 815.900661] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c45537-7cdf-4251-a20a-20c9ac530fb1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.912019] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 815.912504] env[62558]: ERROR oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk due to incomplete transfer. [ 815.912866] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-751e6869-9cdd-469e-ae91-10a207e482bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.923068] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eabccc-21e0-2209-778f-6b72fad20bd1/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 815.923379] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Uploaded image 3d2a9cd1-658b-4ac6-b24b-40043736ee7b to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 815.926160] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 815.926467] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-414f07c9-0833-477a-a053-e2cadb95ca11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.938981] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 815.938981] env[62558]: value = "task-1266686" [ 815.938981] env[62558]: _type = "Task" [ 815.938981] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.950635] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266686, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.025960] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266683, 'name': PowerOffVM_Task, 'duration_secs': 0.250715} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.026266] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.026495] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.026778] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d675e3d-d06f-4eda-bd67-fc791f7a174a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.065152] env[62558]: DEBUG nova.network.neutron [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.087577] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.087870] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.088130] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleting the datastore file [datastore1] ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.088935] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1af4cf9b-fa18-4712-97ea-8e059e4c663f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.097645] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266684, 'name': ReconfigVM_Task, 'duration_secs': 0.380499} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.103795] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff/6d89107f-0727-4b8c-863d-d1e635000bff.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.106020] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-899d0aac-e391-4c7c-8118-c14e050337bf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.109884] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 816.109884] env[62558]: value = "task-1266688" [ 816.109884] env[62558]: _type = "Task" [ 816.109884] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.116462] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266685, 'name': PowerOffVM_Task, 'duration_secs': 0.453899} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.116838] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 816.116838] env[62558]: value = "task-1266689" [ 816.116838] env[62558]: _type = "Task" [ 816.116838] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.119816] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.120023] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.120508] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6273c368-eeaf-486f-ae95-d11bf76758d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.128211] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.131772] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5126a29f-3104-4802-9de6-81bac2c5f064 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.138204] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266689, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.143125] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976e131b-fdb7-4c56-b80b-cba61de96159 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.176955] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a406b75c-1ff1-4f03-80fe-d75610eba1ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.185956] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0e81a3-21fa-4679-8485-a0a5385e23ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.202187] env[62558]: DEBUG nova.compute.provider_tree [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.203651] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.204025] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.204101] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleting the datastore file [datastore2] ebd84b3f-0c5d-40e3-aa70-f8b3054b7109 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.204698] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-557398d7-d6bd-48c7-9984-ffbe13bb86e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.211327] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for the task: (returnval){ [ 816.211327] env[62558]: value = "task-1266691" [ 816.211327] env[62558]: _type = "Task" [ 816.211327] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.221199] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266691, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.261926] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.80698} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.261926] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b/4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 816.262216] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.262556] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e064399e-c09b-4c43-8f81-77e335de67e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.269201] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 816.269201] env[62558]: value = "task-1266692" [ 816.269201] env[62558]: _type = "Task" [ 816.269201] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.281467] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266692, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.297376] env[62558]: INFO nova.compute.manager [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Unrescuing [ 816.297545] env[62558]: DEBUG oslo_concurrency.lockutils [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.297662] env[62558]: DEBUG oslo_concurrency.lockutils [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.297835] env[62558]: DEBUG nova.network.neutron [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.448952] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266686, 'name': Destroy_Task, 'duration_secs': 0.389354} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.448952] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Destroyed the VM [ 816.449281] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 816.449647] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c1e9dc61-dbbd-4bba-95e1-1140bff359e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.457792] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 816.457792] env[62558]: value = "task-1266693" [ 816.457792] env[62558]: _type = "Task" [ 816.457792] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.464861] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266693, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.570263] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.570561] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance network_info: |[{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.571067] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:c8:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7472a75c-48b2-4f52-a2b5-905ae7a9ef3b', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.578359] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating folder: Project (480afedc5a8845e8920815407f4485dc). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.578678] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1e47f7e-b314-4e3f-a2c7-c10c1bfb2843 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.588745] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Created folder: Project (480afedc5a8845e8920815407f4485dc) in parent group-v272451. [ 816.588953] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating folder: Instances. Parent ref: group-v272506. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 816.589218] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88bf1462-e71e-4399-92a8-1f966381504a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.597382] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Created folder: Instances in parent group-v272506. [ 816.597616] env[62558]: DEBUG oslo.service.loopingcall [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.597894] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.598007] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23d02a33-db0b-4d43-822b-05ad6cf9197c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.625968] env[62558]: DEBUG oslo_vmware.api [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446135} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.627269] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.627475] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.627654] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.627831] env[62558]: INFO nova.compute.manager [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Took 1.14 seconds to destroy the instance on the hypervisor. [ 816.628088] env[62558]: DEBUG oslo.service.loopingcall [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.628269] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.628269] env[62558]: value = "task-1266696" [ 816.628269] env[62558]: _type = "Task" [ 816.628269] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.628825] env[62558]: DEBUG nova.compute.manager [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.628921] env[62558]: DEBUG nova.network.neutron [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.636375] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266689, 'name': Rename_Task, 'duration_secs': 0.153687} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.636987] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.637570] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d154ff3f-7111-42e8-bf39-b5378c556056 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.641851] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266696, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.646051] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 816.646051] env[62558]: value = "task-1266697" [ 816.646051] env[62558]: _type = "Task" [ 816.646051] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.653622] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.681224] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.701867] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.702179] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.702346] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.702536] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.702685] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.702835] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.703071] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.704512] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.704512] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.704512] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.704512] env[62558]: DEBUG nova.virt.hardware [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.704734] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbaa3e9-a5b6-411f-9393-ccf7f6582276 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.707964] env[62558]: DEBUG nova.scheduler.client.report [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.719436] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04b4b44-4b4d-4dd0-b9bb-a100425308a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.728104] env[62558]: DEBUG oslo_vmware.api [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Task: {'id': task-1266691, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.411328} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.728933] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.729211] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.729445] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.729668] env[62558]: INFO nova.compute.manager [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Took 1.16 seconds to destroy the instance on the hypervisor. [ 816.730029] env[62558]: DEBUG oslo.service.loopingcall [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.730631] env[62558]: DEBUG nova.compute.manager [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.730743] env[62558]: DEBUG nova.network.neutron [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.779874] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266692, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071752} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.780406] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.781136] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047224e4-4ff9-4cd0-88b0-045a8e726031 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.785819] env[62558]: DEBUG nova.compute.manager [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.786112] env[62558]: DEBUG nova.compute.manager [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing instance network info cache due to event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.786253] env[62558]: DEBUG oslo_concurrency.lockutils [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.786396] env[62558]: DEBUG oslo_concurrency.lockutils [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.786559] env[62558]: DEBUG nova.network.neutron [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.812312] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b/4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.815605] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f43e7f4c-d686-4180-a523-33190055e2aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.836797] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 816.836797] env[62558]: value = "task-1266698" [ 816.836797] env[62558]: _type = "Task" [ 816.836797] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.845980] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266698, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.966952] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266693, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.141434] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266696, 'name': CreateVM_Task, 'duration_secs': 0.410794} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.141614] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.142342] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.142514] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.142852] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.143128] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e32ebbd-4aec-4ef5-8c12-1c16eaa8f6e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.148050] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 817.148050] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52dfd1d9-ba64-9b3f-9124-ce6daf9c8491" [ 817.148050] env[62558]: _type = "Task" [ 817.148050] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.163146] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266697, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.166443] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52dfd1d9-ba64-9b3f-9124-ce6daf9c8491, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.184937] env[62558]: DEBUG nova.compute.manager [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Received event network-vif-plugged-8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.185198] env[62558]: DEBUG oslo_concurrency.lockutils [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] Acquiring lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.185422] env[62558]: DEBUG oslo_concurrency.lockutils [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.185598] env[62558]: DEBUG oslo_concurrency.lockutils [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.185775] env[62558]: DEBUG nova.compute.manager [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] No waiting events found dispatching network-vif-plugged-8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.185995] env[62558]: WARNING nova.compute.manager [req-6a7d1274-aa0e-4990-8ae3-9f12c9584183 req-644e8400-cb25-4207-9cba-c7431d9752a3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Received unexpected event network-vif-plugged-8c7dccfe-46a4-49d4-8622-3e385f26e26c for instance with vm_state building and task_state spawning. [ 817.205103] env[62558]: DEBUG nova.compute.manager [req-f7e9bd8c-cd57-4c1b-b67e-9ddbc14b1168 req-57d05ff9-1819-4559-9bd4-45c94db9f4ad service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Received event network-vif-deleted-2f4de907-3732-4ed6-a741-08fa0ce7db8f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.205103] env[62558]: INFO nova.compute.manager [req-f7e9bd8c-cd57-4c1b-b67e-9ddbc14b1168 req-57d05ff9-1819-4559-9bd4-45c94db9f4ad service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Neutron deleted interface 2f4de907-3732-4ed6-a741-08fa0ce7db8f; detaching it from the instance and deleting it from the info cache [ 817.205103] env[62558]: DEBUG nova.network.neutron [req-f7e9bd8c-cd57-4c1b-b67e-9ddbc14b1168 req-57d05ff9-1819-4559-9bd4-45c94db9f4ad service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.213315] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.213717] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.222648] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.785s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.222648] env[62558]: INFO nova.compute.claims [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.292866] env[62558]: DEBUG nova.network.neutron [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.349399] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266698, 'name': ReconfigVM_Task, 'duration_secs': 0.398053} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.351351] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b/4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.352112] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd63ea59-c0f9-43d2-b132-11613cb3af72 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.359421] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 817.359421] env[62558]: value = "task-1266699" [ 817.359421] env[62558]: _type = "Task" [ 817.359421] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.371612] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266699, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.469628] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266693, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.513800] env[62558]: DEBUG nova.network.neutron [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updated VIF entry in instance network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.514198] env[62558]: DEBUG nova.network.neutron [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.621893] env[62558]: DEBUG nova.network.neutron [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.657515] env[62558]: DEBUG nova.network.neutron [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.664688] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52dfd1d9-ba64-9b3f-9124-ce6daf9c8491, 'name': SearchDatastore_Task, 'duration_secs': 0.021741} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.665234] env[62558]: DEBUG oslo_vmware.api [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266697, 'name': PowerOnVM_Task, 'duration_secs': 0.552892} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.665840] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.665932] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.666175] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.666326] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.666524] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.666913] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.667027] env[62558]: DEBUG nova.compute.manager [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.667306] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6615dd8c-59d1-4c10-b57e-6978839debd7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.669778] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a20445-9f75-48f7-be01-7ad9f9a03422 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.683188] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.683398] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.684679] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-263b6797-0e83-4ace-a8b6-9d8d2b26766b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.690301] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 817.690301] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad36ec-ab7d-15c7-6b09-045a8f94423c" [ 817.690301] env[62558]: _type = "Task" [ 817.690301] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.697385] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad36ec-ab7d-15c7-6b09-045a8f94423c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.699910] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Successfully updated port: 8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.707951] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d260f5cc-e150-4967-aad2-cf7fde882eef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.718065] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6a9e09-27b0-4c0f-80bb-cd98902b88cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.731257] env[62558]: DEBUG nova.compute.utils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.734848] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.735053] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.753719] env[62558]: DEBUG nova.compute.manager [req-f7e9bd8c-cd57-4c1b-b67e-9ddbc14b1168 req-57d05ff9-1819-4559-9bd4-45c94db9f4ad service nova] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Detach interface failed, port_id=2f4de907-3732-4ed6-a741-08fa0ce7db8f, reason: Instance ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 817.777800] env[62558]: DEBUG nova.policy [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.797941] env[62558]: DEBUG oslo_concurrency.lockutils [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.797941] env[62558]: DEBUG nova.objects.instance [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'flavor' on Instance uuid 2ac801d7-af70-46e0-88b3-02caee13497d {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.869414] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266699, 'name': Rename_Task, 'duration_secs': 0.494543} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.869702] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.869989] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e49f69c-5306-4053-9f0c-77703e1c176d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.876035] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 817.876035] env[62558]: value = "task-1266700" [ 817.876035] env[62558]: _type = "Task" [ 817.876035] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.884088] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.968447] env[62558]: DEBUG oslo_vmware.api [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266693, 'name': RemoveSnapshot_Task, 'duration_secs': 1.204387} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.968816] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 817.969022] env[62558]: INFO nova.compute.manager [None req-0d201507-71dd-4943-9c09-cba2c0ef1487 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 17.20 seconds to snapshot the instance on the hypervisor. [ 818.016877] env[62558]: DEBUG oslo_concurrency.lockutils [req-27645cdd-39a2-4903-a2c7-ae430e49bac2 req-beea60f9-ae73-409e-a276-8431e0792d6d service nova] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.040526] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Successfully created port: 8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.125185] env[62558]: INFO nova.compute.manager [-] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Took 1.50 seconds to deallocate network for instance. [ 818.166168] env[62558]: INFO nova.compute.manager [-] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Took 1.44 seconds to deallocate network for instance. [ 818.187686] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.202768] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad36ec-ab7d-15c7-6b09-045a8f94423c, 'name': SearchDatastore_Task, 'duration_secs': 0.027673} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.202768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.202768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.202768] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.203913] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71fe7263-a562-455d-9220-bce48c5cad4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.214022] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 818.214022] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528896c4-25e5-9365-3380-c10c971d1024" [ 818.214022] env[62558]: _type = "Task" [ 818.214022] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.219448] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528896c4-25e5-9365-3380-c10c971d1024, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.236021] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.305962] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8642ba-7f04-47a2-a5f9-ff7b89bf004d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.329935] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.332847] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1012b519-5222-47ae-b4a4-593866b5ccc0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.340205] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 818.340205] env[62558]: value = "task-1266701" [ 818.340205] env[62558]: _type = "Task" [ 818.340205] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.352297] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.390089] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266700, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.613507] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9806da-84d8-464f-aeb5-0d1247c2e9e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.621160] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a357b608-c475-408c-9df3-75fc4fe57bd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.651642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.652733] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc731bf8-ef26-4f1b-b856-01f30510ac9e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.659905] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54bd7ec-799b-4809-bb60-ec199bc502d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.673953] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.674818] env[62558]: DEBUG nova.compute.provider_tree [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.722323] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528896c4-25e5-9365-3380-c10c971d1024, 'name': SearchDatastore_Task, 'duration_secs': 0.011718} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.722323] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.722323] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.722600] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af67a2b4-56cb-489d-95d0-043444991228 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.729294] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 818.729294] env[62558]: value = "task-1266702" [ 818.729294] env[62558]: _type = "Task" [ 818.729294] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.738806] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266702, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.739979] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.841966] env[62558]: DEBUG nova.compute.manager [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Received event network-vif-deleted-e384e059-74fe-48a6-a03e-e6262dee7821 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.842274] env[62558]: DEBUG nova.compute.manager [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Received event network-changed-8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.842452] env[62558]: DEBUG nova.compute.manager [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Refreshing instance network info cache due to event network-changed-8c7dccfe-46a4-49d4-8622-3e385f26e26c. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.842686] env[62558]: DEBUG oslo_concurrency.lockutils [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] Acquiring lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.855196] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266701, 'name': PowerOffVM_Task, 'duration_secs': 0.192374} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.855456] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.860859] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfiguring VM instance instance-00000036 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 818.861262] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dffb5014-eb74-4d14-9923-85f1c924dded {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.884898] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 818.884898] env[62558]: value = "task-1266703" [ 818.884898] env[62558]: _type = "Task" [ 818.884898] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.886108] env[62558]: DEBUG nova.network.neutron [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Updating instance_info_cache with network_info: [{"id": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "address": "fa:16:3e:0b:3c:27", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c7dccfe-46", "ovs_interfaceid": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.899823] env[62558]: DEBUG oslo_vmware.api [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266700, 'name': PowerOnVM_Task, 'duration_secs': 0.719589} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.900967] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.901418] env[62558]: INFO nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Took 9.48 seconds to spawn the instance on the hypervisor. [ 818.901793] env[62558]: DEBUG nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.909871] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a5b20d-cfa0-46a4-88ed-e654ea06ec8f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.915057] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266703, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.177418] env[62558]: DEBUG nova.scheduler.client.report [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.238770] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266702, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471781} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.239072] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.239295] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.239585] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7466e27-aea8-407c-903f-6bc8b6986a32 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.247990] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.248785] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 819.248785] env[62558]: value = "task-1266704" [ 819.248785] env[62558]: _type = "Task" [ 819.248785] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.257732] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.269124] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.269445] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.269583] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.269772] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.269924] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.270670] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.270670] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.270670] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.270670] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.270883] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.270989] env[62558]: DEBUG nova.virt.hardware [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.272081] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6be951d-7a92-46e9-98bf-4d20aba6b20e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.281214] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197f630d-a754-4e39-9834-3168cad654f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.398239] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266703, 'name': ReconfigVM_Task, 'duration_secs': 0.403993} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.398486] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Reconfigured VM instance instance-00000036 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 819.398578] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.399150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.399430] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance network_info: |[{"id": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "address": "fa:16:3e:0b:3c:27", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c7dccfe-46", "ovs_interfaceid": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.399622] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c189c9b7-a24d-4d00-846a-1979fb53d433 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.401322] env[62558]: DEBUG oslo_concurrency.lockutils [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] Acquired lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.401506] env[62558]: DEBUG nova.network.neutron [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Refreshing network info cache for port 8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.402697] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:3c:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c7dccfe-46a4-49d4-8622-3e385f26e26c', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.410925] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating folder: Project (52e1d3e1dcd64e918c1e436e691e290b). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.411431] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab4e1aa4-d634-420d-b588-e8663e147c57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.418364] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 819.418364] env[62558]: value = "task-1266705" [ 819.418364] env[62558]: _type = "Task" [ 819.418364] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.423197] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created folder: Project (52e1d3e1dcd64e918c1e436e691e290b) in parent group-v272451. [ 819.423402] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating folder: Instances. Parent ref: group-v272509. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.426196] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ddf6316-26af-4da2-a5c0-6412a93faa04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.431539] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.435828] env[62558]: INFO nova.compute.manager [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Took 38.79 seconds to build instance. [ 819.441398] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created folder: Instances in parent group-v272509. [ 819.441638] env[62558]: DEBUG oslo.service.loopingcall [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.441870] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.442179] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a911729-a5d6-4925-9eac-d88b5e0925e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.463647] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.463647] env[62558]: value = "task-1266708" [ 819.463647] env[62558]: _type = "Task" [ 819.463647] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.471332] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.531448] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "6d89107f-0727-4b8c-863d-d1e635000bff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.531448] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.531448] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.531691] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.532778] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.534304] env[62558]: INFO nova.compute.manager [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Terminating instance [ 819.536325] env[62558]: DEBUG nova.compute.manager [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 819.536531] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 819.537363] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba1aa83-0529-4b5b-9f1a-b6dc51b0c4f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.545122] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 819.546584] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e0a83e5-5941-4e98-a015-dcfb4f5be105 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.551453] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 819.551453] env[62558]: value = "task-1266709" [ 819.551453] env[62558]: _type = "Task" [ 819.551453] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.559779] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.560778] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Successfully updated port: 8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.686945] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.686945] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.693928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.339s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.698399] env[62558]: INFO nova.compute.claims [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.762108] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076517} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.762491] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.763329] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a6e506-7b91-41d3-9f9c-93a40db6dc11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.790845] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.791169] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acb443e5-128c-4c0f-b6ca-047e38c7fc36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.813020] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 819.813020] env[62558]: value = "task-1266710" [ 819.813020] env[62558]: _type = "Task" [ 819.813020] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.822642] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.929387] env[62558]: DEBUG oslo_vmware.api [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266705, 'name': PowerOnVM_Task, 'duration_secs': 0.432921} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.929697] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 819.929959] env[62558]: DEBUG nova.compute.manager [None req-728bc86b-8ef2-4d9c-9c6c-628d39449514 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.930808] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d5e32d-b2e9-4e23-b8d3-99e199a9623f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.941840] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a4fae36c-ebfa-4c0d-92e3-899644a726b0 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.691s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.980156] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.070688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.071537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.071731] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.073085] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266709, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.193638] env[62558]: DEBUG nova.compute.utils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.195893] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.196083] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.213734] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.213975] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.323015] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.446505] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.460746] env[62558]: DEBUG nova.policy [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.480299] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.561956] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266709, 'name': PowerOffVM_Task, 'duration_secs': 0.538775} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.562898] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.563188] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.563458] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79158947-cd81-41b8-a599-8310cf382f68 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.613076] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.625668] env[62558]: DEBUG nova.network.neutron [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Updated VIF entry in instance network info cache for port 8c7dccfe-46a4-49d4-8622-3e385f26e26c. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.626071] env[62558]: DEBUG nova.network.neutron [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Updating instance_info_cache with network_info: [{"id": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "address": "fa:16:3e:0b:3c:27", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c7dccfe-46", "ovs_interfaceid": "8c7dccfe-46a4-49d4-8622-3e385f26e26c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.697757] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.791785] env[62558]: DEBUG nova.network.neutron [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Updating instance_info_cache with network_info: [{"id": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "address": "fa:16:3e:87:40:76", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e3c1064-af", "ovs_interfaceid": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.825681] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266710, 'name': ReconfigVM_Task, 'duration_secs': 0.919047} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.825969] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.826639] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12480556-0737-461d-a978-4cfd3f68ccd7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.834846] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 820.834846] env[62558]: value = "task-1266712" [ 820.834846] env[62558]: _type = "Task" [ 820.834846] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.843294] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266712, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.916528] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Successfully created port: 9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.920995] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Received event network-vif-plugged-8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.920995] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Acquiring lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.920995] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.921638] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.921638] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] No waiting events found dispatching network-vif-plugged-8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.921638] env[62558]: WARNING nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Received unexpected event network-vif-plugged-8e3c1064-af85-4001-9ba5-1ebba633bbd8 for instance with vm_state building and task_state spawning. [ 820.921638] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Received event network-changed-8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.923049] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Refreshing instance network info cache due to event network-changed-8e3c1064-af85-4001-9ba5-1ebba633bbd8. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.923049] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Acquiring lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.977407] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.980919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.029864] env[62558]: DEBUG nova.compute.manager [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.030815] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e723fb2b-12e1-4da2-b9ad-3d0b8dbf31ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.126336] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47285a97-3a73-4f8d-9c6d-12619ade4db9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.131042] env[62558]: DEBUG oslo_concurrency.lockutils [req-a71c4e38-2a54-42e7-a5a6-fcca8967c38f req-44c38ae9-8cc5-4ff1-b2e4-0b06ee53c141 service nova] Releasing lock "refresh_cache-4a3f7642-1b9c-40d8-973b-5153b559bda0" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.137341] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d52cf6-840f-434d-9295-bc8173de6e0f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.173572] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05fb2ad-6044-4631-80c3-5bf75503fb4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.182235] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49563de2-4413-4714-b7c6-20a1934566b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.199750] env[62558]: DEBUG nova.compute.provider_tree [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.294271] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.294827] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Instance network_info: |[{"id": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "address": "fa:16:3e:87:40:76", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e3c1064-af", "ovs_interfaceid": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 821.295786] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Acquired lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.295786] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Refreshing network info cache for port 8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.297369] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:40:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e3c1064-af85-4001-9ba5-1ebba633bbd8', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.305211] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating folder: Project (c184479dcbc849ea983347809d5fc3b2). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.305809] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40f3dcc4-ac76-4b7b-93d2-2a0b6cb42c0a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.318048] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created folder: Project (c184479dcbc849ea983347809d5fc3b2) in parent group-v272451. [ 821.318273] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating folder: Instances. Parent ref: group-v272512. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.318506] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47e3c598-010e-4077-94bb-51622262a9b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.327910] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created folder: Instances in parent group-v272512. [ 821.328125] env[62558]: DEBUG oslo.service.loopingcall [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.328317] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.328550] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81a6f322-82ac-4ec4-ab5e-5a7f7e07ba95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.353409] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266712, 'name': Rename_Task, 'duration_secs': 0.193754} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.354885] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.355181] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.355181] env[62558]: value = "task-1266715" [ 821.355181] env[62558]: _type = "Task" [ 821.355181] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.355414] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85fbc2b6-1fb6-4330-a8c8-a452fe4e90da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.365699] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266715, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.366687] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 821.366687] env[62558]: value = "task-1266716" [ 821.366687] env[62558]: _type = "Task" [ 821.366687] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.374158] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.419020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.419020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.419020] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleting the datastore file [datastore2] 6d89107f-0727-4b8c-863d-d1e635000bff {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.419020] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8975261c-6614-4a54-bb49-1dea86472814 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.424677] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 821.424677] env[62558]: value = "task-1266717" [ 821.424677] env[62558]: _type = "Task" [ 821.424677] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.435046] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266717, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.480049] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.544461] env[62558]: INFO nova.compute.manager [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] instance snapshotting [ 821.547681] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c376de4-5922-4c9c-b622-01661770f864 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.572510] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170f8ff7-1494-4c74-a98e-386b1ec4b120 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.709095] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.730518] env[62558]: ERROR nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [req-6c90eb5a-cfe3-4c3c-a65d-b7e18e08a0b7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6c90eb5a-cfe3-4c3c-a65d-b7e18e08a0b7"}]} [ 821.735193] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.735479] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.735638] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.735880] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.736055] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.736212] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.736423] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.736586] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.736757] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.736923] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.737119] env[62558]: DEBUG nova.virt.hardware [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.738921] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910e8bf4-72fd-4138-ae9a-644f7366710a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.747647] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51262670-8077-4fd0-b754-c40ee5fbeaba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.754012] env[62558]: DEBUG nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 821.777928] env[62558]: DEBUG nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 821.778194] env[62558]: DEBUG nova.compute.provider_tree [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.790578] env[62558]: DEBUG nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 821.811976] env[62558]: DEBUG nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 821.868122] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266715, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.883378] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266716, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.934929] env[62558]: DEBUG oslo_vmware.api [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266717, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.36371} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.937679] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.937873] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.938072] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.938258] env[62558]: INFO nova.compute.manager [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Took 2.40 seconds to destroy the instance on the hypervisor. [ 821.938579] env[62558]: DEBUG oslo.service.loopingcall [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.939050] env[62558]: DEBUG nova.compute.manager [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.939174] env[62558]: DEBUG nova.network.neutron [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.981791] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266708, 'name': CreateVM_Task, 'duration_secs': 2.037149} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.984707] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.986029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.986029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.986426] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.986671] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce2df800-96e3-4607-99ba-de0605fc5338 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.994646] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 821.994646] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a51ee-08ca-4406-087a-b79904e5b6fb" [ 821.994646] env[62558]: _type = "Task" [ 821.994646] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.006276] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a51ee-08ca-4406-087a-b79904e5b6fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.083952] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 822.084102] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0a8339d2-37e6-48a6-9899-b61c97b5e778 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.086772] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Updated VIF entry in instance network info cache for port 8e3c1064-af85-4001-9ba5-1ebba633bbd8. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.087115] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Updating instance_info_cache with network_info: [{"id": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "address": "fa:16:3e:87:40:76", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e3c1064-af", "ovs_interfaceid": "8e3c1064-af85-4001-9ba5-1ebba633bbd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.095918] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 822.095918] env[62558]: value = "task-1266718" [ 822.095918] env[62558]: _type = "Task" [ 822.095918] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.106534] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266718, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.195291] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba98618-f358-4f5a-bf5f-035434d5275e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.211023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e8eeb3-8392-4978-98d9-b41dc1cf8e9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.245956] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1092107c-50b3-4527-b3bd-22cb32ef1025 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.254993] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a321a99c-a23a-4d6e-a8b0-5f994eda874f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.270020] env[62558]: DEBUG nova.compute.provider_tree [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.368529] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266715, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.378460] env[62558]: DEBUG oslo_vmware.api [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266716, 'name': PowerOnVM_Task, 'duration_secs': 0.818733} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.378876] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.379213] env[62558]: INFO nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Took 8.32 seconds to spawn the instance on the hypervisor. [ 822.379551] env[62558]: DEBUG nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.380452] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a163eeb6-c231-4fcf-8536-a4c7c9bc947f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.509608] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a51ee-08ca-4406-087a-b79904e5b6fb, 'name': SearchDatastore_Task, 'duration_secs': 0.013278} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.510148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.510513] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.510778] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.510955] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.511181] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.511549] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-818452d3-3109-4944-8527-4b8638f68079 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.520339] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.520523] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.521276] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a43ab3df-b87a-451a-b0a3-b13e2c2fd212 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.526269] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 822.526269] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524dff40-9cb3-a8c8-a0f7-e2a3a8bd5bb7" [ 822.526269] env[62558]: _type = "Task" [ 822.526269] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.536318] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524dff40-9cb3-a8c8-a0f7-e2a3a8bd5bb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.591687] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Releasing lock "refresh_cache-e846e43d-68ec-4de3-ba62-e538643b6e4b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.592048] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.592571] env[62558]: DEBUG nova.compute.manager [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing instance network info cache due to event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.592812] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Acquiring lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.593020] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Acquired lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.593596] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing network info cache for port 6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.606660] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266718, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.672625] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Successfully updated port: 9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.715680] env[62558]: DEBUG nova.network.neutron [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.774652] env[62558]: DEBUG nova.scheduler.client.report [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.868938] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266715, 'name': CreateVM_Task, 'duration_secs': 1.435505} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.869728] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.869941] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.870090] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.872649] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.872649] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baa17f81-1af9-4e6c-8379-1b3b48dfd495 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.875441] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 822.875441] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52600a58-0550-23f8-7796-2574363f3706" [ 822.875441] env[62558]: _type = "Task" [ 822.875441] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.883951] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52600a58-0550-23f8-7796-2574363f3706, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.903864] env[62558]: INFO nova.compute.manager [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Took 34.61 seconds to build instance. [ 822.951812] env[62558]: DEBUG nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Received event network-vif-deleted-cc1b2699-3507-45b0-afe0-572fd6966fdf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.952133] env[62558]: DEBUG nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-vif-plugged-9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.952455] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.952782] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.952979] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.953178] env[62558]: DEBUG nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] No waiting events found dispatching network-vif-plugged-9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.953344] env[62558]: WARNING nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received unexpected event network-vif-plugged-9d645945-9804-4520-bd40-e7475a0dd042 for instance with vm_state building and task_state spawning. [ 822.953510] env[62558]: DEBUG nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-changed-9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.953664] env[62558]: DEBUG nova.compute.manager [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing instance network info cache due to event network-changed-9d645945-9804-4520-bd40-e7475a0dd042. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.953852] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.953991] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.954167] env[62558]: DEBUG nova.network.neutron [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing network info cache for port 9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.036877] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524dff40-9cb3-a8c8-a0f7-e2a3a8bd5bb7, 'name': SearchDatastore_Task, 'duration_secs': 0.015505} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.037713] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f84344d-28e2-4875-8559-2461e63db17d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.046102] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 823.046102] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52956def-1de6-dc25-09cf-150f34e72091" [ 823.046102] env[62558]: _type = "Task" [ 823.046102] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.054520] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52956def-1de6-dc25-09cf-150f34e72091, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.107494] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266718, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.177098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.218089] env[62558]: INFO nova.compute.manager [-] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Took 1.28 seconds to deallocate network for instance. [ 823.280142] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.588s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.280684] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.283191] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.658s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.286854] env[62558]: DEBUG nova.objects.instance [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lazy-loading 'resources' on Instance uuid 3c8c2d25-981e-49de-b5c0-e0d12dd91378 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.387216] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52600a58-0550-23f8-7796-2574363f3706, 'name': SearchDatastore_Task, 'duration_secs': 0.011335} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.387551] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.387737] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.387946] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.399017] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updated VIF entry in instance network info cache for port 6fcb24d4-f48e-4628-995d-10f53f01154b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.399396] env[62558]: DEBUG nova.network.neutron [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [{"id": "6fcb24d4-f48e-4628-995d-10f53f01154b", "address": "fa:16:3e:09:d7:9e", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb24d4-f4", "ovs_interfaceid": "6fcb24d4-f48e-4628-995d-10f53f01154b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.407427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-589a41e5-321c-4160-ad63-3b41017bdcea tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.659s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.503601] env[62558]: DEBUG nova.network.neutron [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.558597] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52956def-1de6-dc25-09cf-150f34e72091, 'name': SearchDatastore_Task, 'duration_secs': 0.015076} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.558984] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.559274] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4a3f7642-1b9c-40d8-973b-5153b559bda0/4a3f7642-1b9c-40d8-973b-5153b559bda0.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.559548] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.559738] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.560125] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35c81087-c53b-46e8-a418-407de9958f17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.561957] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01fc5cb2-a1af-46e6-92db-605c105fbe29 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.569160] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 823.569160] env[62558]: value = "task-1266719" [ 823.569160] env[62558]: _type = "Task" [ 823.569160] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.576502] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.576502] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.576980] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8022c9a7-2bb4-4bc7-a8d1-1110057884a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.582174] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.586090] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 823.586090] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bbd04d-5b34-7731-3ce6-be1ef406b8b1" [ 823.586090] env[62558]: _type = "Task" [ 823.586090] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.597215] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bbd04d-5b34-7731-3ce6-be1ef406b8b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009095} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.597956] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e297dfe3-c2bc-470d-96d1-2be65e8b198e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.611625] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 823.611625] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d006e9-8adb-43a2-1df0-b45d8c8c8613" [ 823.611625] env[62558]: _type = "Task" [ 823.611625] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.612237] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266718, 'name': CreateSnapshot_Task, 'duration_secs': 1.042627} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.612736] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 823.616848] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6e1082-6bab-4092-8c90-75e58ddf5a3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.620373] env[62558]: DEBUG nova.network.neutron [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.626331] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d006e9-8adb-43a2-1df0-b45d8c8c8613, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.629388] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.629632] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] e846e43d-68ec-4de3-ba62-e538643b6e4b/e846e43d-68ec-4de3-ba62-e538643b6e4b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.635595] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ed1dcda-427d-4128-a9df-4712a4a953f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.639021] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 823.639021] env[62558]: value = "task-1266720" [ 823.639021] env[62558]: _type = "Task" [ 823.639021] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.646701] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266720, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.725224] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.788921] env[62558]: DEBUG nova.compute.utils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.794796] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.795199] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.884754] env[62558]: DEBUG nova.policy [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9ea7dcf145f4d4098ee057593c1ebe7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7663e884862d43b49eabe6396d8cd107', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.902447] env[62558]: DEBUG oslo_concurrency.lockutils [req-14f80832-9185-48b7-b9a4-bb02e7457f03 req-96755a4f-f113-49f5-9a49-84b93e802a29 service nova] Releasing lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.912468] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.080698] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266719, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.125719] env[62558]: DEBUG oslo_concurrency.lockutils [req-e347f6d2-6cf2-413c-8ae3-bcb9b56d051d req-0bfa40e0-660b-41c1-bf73-5cdbf4e08cf9 service nova] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.126498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.126792] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.144194] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 824.148333] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1325731c-3b93-4d67-9615-a9123f0629d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.170026] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266720, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.172161] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 824.172161] env[62558]: value = "task-1266721" [ 824.172161] env[62558]: _type = "Task" [ 824.172161] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.184507] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266721, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.270049] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f578df0-2c73-47d7-829d-bce2306b9d8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.279660] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56305624-dddd-412d-ace0-e2d8726602ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.321616] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.325625] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd73c3aa-b751-4a23-b5ab-8e23fdc39751 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.338029] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7076714-c701-4c62-9281-77651fc534f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.351648] env[62558]: DEBUG nova.compute.provider_tree [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.430592] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Successfully created port: 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.439846] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.470378] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.470569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.581344] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598457} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.581829] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4a3f7642-1b9c-40d8-973b-5153b559bda0/4a3f7642-1b9c-40d8-973b-5153b559bda0.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.582230] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.582600] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04b8af18-4a8f-4e7c-a074-6531ef7743d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.591380] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 824.591380] env[62558]: value = "task-1266722" [ 824.591380] env[62558]: _type = "Task" [ 824.591380] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.602021] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266722, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.660098] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266720, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.761512} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.660098] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] e846e43d-68ec-4de3-ba62-e538643b6e4b/e846e43d-68ec-4de3-ba62-e538643b6e4b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.660625] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.662231] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2dc80d60-0885-4bff-90f3-556551ffc54b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.668114] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 824.668114] env[62558]: value = "task-1266723" [ 824.668114] env[62558]: _type = "Task" [ 824.668114] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.679750] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.680830] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.687785] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266721, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.823406] env[62558]: DEBUG nova.network.neutron [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.873954] env[62558]: ERROR nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] [req-844a263c-2d43-487c-9f8e-b5dae6e2bdd0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-844a263c-2d43-487c-9f8e-b5dae6e2bdd0"}]} [ 824.889625] env[62558]: DEBUG nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 824.903455] env[62558]: DEBUG nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 824.904103] env[62558]: DEBUG nova.compute.provider_tree [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.919279] env[62558]: DEBUG nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 824.942378] env[62558]: DEBUG nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 824.985130] env[62558]: DEBUG nova.compute.manager [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.985130] env[62558]: DEBUG nova.compute.manager [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing instance network info cache due to event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.985130] env[62558]: DEBUG oslo_concurrency.lockutils [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.985130] env[62558]: DEBUG oslo_concurrency.lockutils [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.985130] env[62558]: DEBUG nova.network.neutron [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.101413] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266722, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072403} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.104031] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.105175] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f308ce8-b0b8-419a-8a75-9453d8efa4c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.130341] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 4a3f7642-1b9c-40d8-973b-5153b559bda0/4a3f7642-1b9c-40d8-973b-5153b559bda0.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.133236] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b38dcbd-dda5-4679-ab73-8062e3bd6be6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.153032] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 825.153032] env[62558]: value = "task-1266724" [ 825.153032] env[62558]: _type = "Task" [ 825.153032] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.163762] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.179827] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.261073} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.182618] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.186936] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423c651b-90e3-48ce-ba4a-91a078ae89ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.189099] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266721, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.210617] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] e846e43d-68ec-4de3-ba62-e538643b6e4b/e846e43d-68ec-4de3-ba62-e538643b6e4b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.213398] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6c4a5d7-84f9-4459-be9a-d8f55685e06a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.232705] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 825.232705] env[62558]: value = "task-1266725" [ 825.232705] env[62558]: _type = "Task" [ 825.232705] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.242826] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266725, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.326517] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.327264] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance network_info: |[{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.327526] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:6f:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d645945-9804-4520-bd40-e7475a0dd042', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.334922] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating folder: Project (bd56f6c392f14aa880ef4891990c1d44). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.337933] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.339596] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8a0a573d-b1bc-4c32-8510-27cf705d15c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.350608] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created folder: Project (bd56f6c392f14aa880ef4891990c1d44) in parent group-v272451. [ 825.350829] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating folder: Instances. Parent ref: group-v272517. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.353380] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c52abcbf-0488-416d-9162-d1c758116a30 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.363492] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.363632] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.363795] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.363983] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.364165] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.364319] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.364525] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.364687] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.364856] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.365029] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.365214] env[62558]: DEBUG nova.virt.hardware [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.365994] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cfe709-dc47-44f4-a852-18233541736e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.371538] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created folder: Instances in parent group-v272517. [ 825.371791] env[62558]: DEBUG oslo.service.loopingcall [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.372569] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.372835] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ff26b73-92ad-4de9-bba6-6e76f82805fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.392420] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0fc79c-3754-4000-a6fd-49a34d9534ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.396369] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d7196e-4728-4b8c-b7fe-1774df85a96d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.402748] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.402748] env[62558]: value = "task-1266728" [ 825.402748] env[62558]: _type = "Task" [ 825.402748] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.417297] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7238ec97-3054-486d-a695-0a9be2a1beb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.424322] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266728, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.456776] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871e4e9c-2973-435d-9484-0bfd7c9e118b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.465691] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87bee70-f6af-415f-bd47-eb973bb205ec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.482036] env[62558]: DEBUG nova.compute.provider_tree [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.662702] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.685300] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266721, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.714161] env[62558]: DEBUG nova.network.neutron [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updated VIF entry in instance network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.714606] env[62558]: DEBUG nova.network.neutron [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.743487] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266725, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.913478] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266728, 'name': CreateVM_Task, 'duration_secs': 0.474139} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.913689] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.914464] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.914642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.914979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.915356] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f43e3e67-de41-49f4-bafb-e969612bcdb5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.920191] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 825.920191] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b153a1-c07d-7f84-6c66-5e543cbd0127" [ 825.920191] env[62558]: _type = "Task" [ 825.920191] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.928448] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b153a1-c07d-7f84-6c66-5e543cbd0127, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.971307] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Successfully updated port: 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.014302] env[62558]: DEBUG nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 826.014658] env[62558]: DEBUG nova.compute.provider_tree [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 89 to 90 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 826.014858] env[62558]: DEBUG nova.compute.provider_tree [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.163574] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.183467] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266721, 'name': CloneVM_Task, 'duration_secs': 1.987199} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.183744] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Created linked-clone VM from snapshot [ 826.184465] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464e48d2-7330-41de-8e8e-d9b8dacc7cab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.191309] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Uploading image d7879362-6b66-4ea0-a511-1ecc21020cf6 {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 826.212290] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 826.212290] env[62558]: value = "vm-272516" [ 826.212290] env[62558]: _type = "VirtualMachine" [ 826.212290] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 826.212574] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3aa3438f-26e5-4109-b817-53e3288f51a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.216822] env[62558]: DEBUG oslo_concurrency.lockutils [req-cb73d24a-d64e-4420-9f44-6cdd5165939b req-0caed74a-116b-4f7b-9baf-5f6b01e84b55 service nova] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.219617] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lease: (returnval){ [ 826.219617] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d59dd1-a9d4-7896-ffcf-78c3c22c8ec2" [ 826.219617] env[62558]: _type = "HttpNfcLease" [ 826.219617] env[62558]: } obtained for exporting VM: (result){ [ 826.219617] env[62558]: value = "vm-272516" [ 826.219617] env[62558]: _type = "VirtualMachine" [ 826.219617] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 826.219964] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the lease: (returnval){ [ 826.219964] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d59dd1-a9d4-7896-ffcf-78c3c22c8ec2" [ 826.219964] env[62558]: _type = "HttpNfcLease" [ 826.219964] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 826.226348] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 826.226348] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d59dd1-a9d4-7896-ffcf-78c3c22c8ec2" [ 826.226348] env[62558]: _type = "HttpNfcLease" [ 826.226348] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 826.242825] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266725, 'name': ReconfigVM_Task, 'duration_secs': 0.605538} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.243105] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Reconfigured VM instance instance-00000042 to attach disk [datastore2] e846e43d-68ec-4de3-ba62-e538643b6e4b/e846e43d-68ec-4de3-ba62-e538643b6e4b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.243813] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0a237de-3858-4958-9283-958b15baea8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.250186] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 826.250186] env[62558]: value = "task-1266730" [ 826.250186] env[62558]: _type = "Task" [ 826.250186] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.259544] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266730, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.431572] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b153a1-c07d-7f84-6c66-5e543cbd0127, 'name': SearchDatastore_Task, 'duration_secs': 0.011979} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.431991] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.432365] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.432673] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.432866] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.433146] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.433441] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b087a76c-7643-4f7e-b1f6-7396ef4e559d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.441937] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.442198] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.442957] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9e2f947-0e29-4883-bfb5-38a907662c51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.448077] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 826.448077] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528eddbf-af7a-bc2e-c880-09bfe549acf1" [ 826.448077] env[62558]: _type = "Task" [ 826.448077] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.455870] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528eddbf-af7a-bc2e-c880-09bfe549acf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.473666] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.473851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.474016] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.520180] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.237s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.522695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.025s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.524300] env[62558]: INFO nova.compute.claims [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.545463] env[62558]: INFO nova.scheduler.client.report [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Deleted allocations for instance 3c8c2d25-981e-49de-b5c0-e0d12dd91378 [ 826.665060] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266724, 'name': ReconfigVM_Task, 'duration_secs': 1.376426} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.665332] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 4a3f7642-1b9c-40d8-973b-5153b559bda0/4a3f7642-1b9c-40d8-973b-5153b559bda0.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.665996] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c527727a-b46c-4864-9dd4-02cf08ad815c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.672560] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 826.672560] env[62558]: value = "task-1266731" [ 826.672560] env[62558]: _type = "Task" [ 826.672560] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.680256] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266731, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.730369] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 826.730369] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d59dd1-a9d4-7896-ffcf-78c3c22c8ec2" [ 826.730369] env[62558]: _type = "HttpNfcLease" [ 826.730369] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 826.730601] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 826.730601] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d59dd1-a9d4-7896-ffcf-78c3c22c8ec2" [ 826.730601] env[62558]: _type = "HttpNfcLease" [ 826.730601] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 826.731606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796f8d9a-0383-4812-9b0e-a8ad0f8f6daa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.739790] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 826.739983] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 826.804940] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266730, 'name': Rename_Task, 'duration_secs': 0.154757} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.805502] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.805765] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26d6fc8a-78c1-4f8c-9c16-271484ea28da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.811623] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 826.811623] env[62558]: value = "task-1266732" [ 826.811623] env[62558]: _type = "Task" [ 826.811623] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.818816] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266732, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.832514] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5b4a2c01-a0ea-4ef4-90a9-a59d17fd736a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.958957] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528eddbf-af7a-bc2e-c880-09bfe549acf1, 'name': SearchDatastore_Task, 'duration_secs': 0.009322} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.959804] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e6a56b3-56e0-4036-b706-1e9aa5ba9d5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.965142] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 826.965142] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5263062f-56e7-4238-258d-4fee673a3ffc" [ 826.965142] env[62558]: _type = "Task" [ 826.965142] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.972788] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5263062f-56e7-4238-258d-4fee673a3ffc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.023588] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.028136] env[62558]: DEBUG nova.compute.manager [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.028366] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.028582] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.028753] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.028943] env[62558]: DEBUG nova.compute.manager [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] No waiting events found dispatching network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.029128] env[62558]: WARNING nova.compute.manager [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received unexpected event network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a for instance with vm_state building and task_state spawning. [ 827.029292] env[62558]: DEBUG nova.compute.manager [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.029469] env[62558]: DEBUG nova.compute.manager [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing instance network info cache due to event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.029642] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.053341] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c1274f7-c800-4055-96b1-5cc2001bf8f8 tempest-ServerMetadataTestJSON-1020929201 tempest-ServerMetadataTestJSON-1020929201-project-member] Lock "3c8c2d25-981e-49de-b5c0-e0d12dd91378" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.369s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.184699] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266731, 'name': Rename_Task, 'duration_secs': 0.258525} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.185676] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.185676] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-294e5de2-c121-4f12-ae0b-770d1f6efb19 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.194914] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 827.194914] env[62558]: value = "task-1266733" [ 827.194914] env[62558]: _type = "Task" [ 827.194914] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.204819] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.205893] env[62558]: DEBUG nova.network.neutron [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.323617] env[62558]: DEBUG oslo_vmware.api [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266732, 'name': PowerOnVM_Task, 'duration_secs': 0.463835} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.324066] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.324506] env[62558]: INFO nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Took 8.08 seconds to spawn the instance on the hypervisor. [ 827.324760] env[62558]: DEBUG nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.325758] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b5da3a-9ebf-427e-86b0-d45f0d3caf09 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.477716] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5263062f-56e7-4238-258d-4fee673a3ffc, 'name': SearchDatastore_Task, 'duration_secs': 0.009353} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.477716] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.477716] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 4526df2a-f801-4de8-8218-497a7d22034f/4526df2a-f801-4de8-8218-497a7d22034f.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.477716] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-430c89ad-219c-4671-b5b1-40ce36a27fb6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.484556] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 827.484556] env[62558]: value = "task-1266734" [ 827.484556] env[62558]: _type = "Task" [ 827.484556] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.493096] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.711658] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.712095] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance network_info: |[{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.712687] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266733, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.712816] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.713244] env[62558]: DEBUG nova.network.neutron [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.717609] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:85:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53ebf5df-5ecb-4a0c-a163-d88165639de0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.723886] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating folder: Project (7663e884862d43b49eabe6396d8cd107). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.728340] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30843d63-a73c-4c2a-81a7-8e858869b6d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.740570] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created folder: Project (7663e884862d43b49eabe6396d8cd107) in parent group-v272451. [ 827.740790] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating folder: Instances. Parent ref: group-v272520. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.741124] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05d6735f-0454-4aee-a00e-b46a32dd0914 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.754291] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created folder: Instances in parent group-v272520. [ 827.756044] env[62558]: DEBUG oslo.service.loopingcall [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.756044] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.756044] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00d0e087-d5ab-4fdc-b306-5628c1d94380 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.780348] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.780348] env[62558]: value = "task-1266737" [ 827.780348] env[62558]: _type = "Task" [ 827.780348] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.790486] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266737, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.854153] env[62558]: INFO nova.compute.manager [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Took 27.79 seconds to build instance. [ 828.000265] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266734, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.054055] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d892dbd9-0dd2-4fe6-8e45-3ec67090aff6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.062116] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e0fa23-4f61-4f26-b67c-4f34c5bb6b72 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.105076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22353d9-20a6-46a4-b3ff-3141e325a9cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.116181] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2c8823-3981-43b4-be30-fa79fa541bf2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.133467] env[62558]: DEBUG nova.compute.provider_tree [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.207553] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266733, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.292151] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266737, 'name': CreateVM_Task, 'duration_secs': 0.37637} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.292360] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.293219] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.293458] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.293923] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.294373] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b4b6246-076e-47f2-aeb1-e233163716d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.299699] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 828.299699] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a9feb4-6837-b204-f0bc-72bdea4e4596" [ 828.299699] env[62558]: _type = "Task" [ 828.299699] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.308172] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a9feb4-6837-b204-f0bc-72bdea4e4596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.356793] env[62558]: DEBUG oslo_concurrency.lockutils [None req-934ce704-7387-41db-8d44-18ca7373ed3e tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.500s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.501265] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696495} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.501711] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 4526df2a-f801-4de8-8218-497a7d22034f/4526df2a-f801-4de8-8218-497a7d22034f.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.501937] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.502264] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a2e5ab6-5417-43a1-ae58-3f77c45551f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.510407] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 828.510407] env[62558]: value = "task-1266738" [ 828.510407] env[62558]: _type = "Task" [ 828.510407] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.519729] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.618748] env[62558]: DEBUG nova.network.neutron [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updated VIF entry in instance network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.622161] env[62558]: DEBUG nova.network.neutron [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.639972] env[62558]: DEBUG nova.scheduler.client.report [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.712973] env[62558]: DEBUG oslo_vmware.api [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266733, 'name': PowerOnVM_Task, 'duration_secs': 1.233366} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.713735] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.714056] env[62558]: INFO nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Took 12.03 seconds to spawn the instance on the hypervisor. [ 828.714330] env[62558]: DEBUG nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.715584] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c56aa26-8df5-4d44-9c05-2ce186fc2cf4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.814300] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a9feb4-6837-b204-f0bc-72bdea4e4596, 'name': SearchDatastore_Task, 'duration_secs': 0.021938} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.814626] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.816098] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.816098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.816098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.816098] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.816098] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c50d4ab-e702-4b40-a10a-6bbeafd08b9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.828120] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.828338] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.829173] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93ddbf77-cfa9-4a87-8d69-b0ef58191d1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.836913] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 828.836913] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52000ec7-4f2b-5073-18ca-63ba184d79a1" [ 828.836913] env[62558]: _type = "Task" [ 828.836913] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.845646] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52000ec7-4f2b-5073-18ca-63ba184d79a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.859425] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.021230] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131641} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.021534] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.022440] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d422ff36-8f19-4816-8197-6ca94449a006 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.046621] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 4526df2a-f801-4de8-8218-497a7d22034f/4526df2a-f801-4de8-8218-497a7d22034f.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.046816] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc97e154-1ea3-4eef-8960-6a84d2a25e3e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.069964] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 829.069964] env[62558]: value = "task-1266739" [ 829.069964] env[62558]: _type = "Task" [ 829.069964] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.079795] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266739, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.123377] env[62558]: DEBUG oslo_concurrency.lockutils [req-821912ac-3ebc-4378-8b98-2719c470410e req-9f18547e-89f0-4fcc-9025-121d81cc7c63 service nova] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.144174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.144708] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.147381] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.484s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.148890] env[62558]: INFO nova.compute.claims [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.246905] env[62558]: INFO nova.compute.manager [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Took 31.82 seconds to build instance. [ 829.348740] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52000ec7-4f2b-5073-18ca-63ba184d79a1, 'name': SearchDatastore_Task, 'duration_secs': 0.015406} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.349580] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5029cd3f-25ec-4274-9bd6-e416e048a463 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.356738] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 829.356738] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ce794e-fc3d-39ab-aa2f-03f42a300ee9" [ 829.356738] env[62558]: _type = "Task" [ 829.356738] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.371210] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ce794e-fc3d-39ab-aa2f-03f42a300ee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.389652] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.447047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.447217] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.447471] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.447668] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.447845] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.451450] env[62558]: INFO nova.compute.manager [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Terminating instance [ 829.452986] env[62558]: DEBUG nova.compute.manager [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.453108] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.454311] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f585fe-237e-4b08-9a9d-f21e8b159539 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.464255] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.464842] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8349b58a-ccc0-44ca-a724-6b02cb7b95d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.472037] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 829.472037] env[62558]: value = "task-1266740" [ 829.472037] env[62558]: _type = "Task" [ 829.472037] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.480485] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.580462] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.656197] env[62558]: DEBUG nova.compute.utils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.658611] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.658802] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.711618] env[62558]: DEBUG nova.policy [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.749503] env[62558]: DEBUG oslo_concurrency.lockutils [None req-189b6297-0dd2-4256-82fa-fc30c3971452 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.052s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.870548] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ce794e-fc3d-39ab-aa2f-03f42a300ee9, 'name': SearchDatastore_Task, 'duration_secs': 0.009909} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.870865] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.871209] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.871520] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03ff4464-b57e-4b40-9392-78c14d69a131 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.878588] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 829.878588] env[62558]: value = "task-1266741" [ 829.878588] env[62558]: _type = "Task" [ 829.878588] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.886839] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.983016] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266740, 'name': PowerOffVM_Task, 'duration_secs': 0.237749} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.983336] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.983511] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.983776] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d333bfb-2be5-4dd1-87a5-29b22da931a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.032237] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Successfully created port: 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.051806] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.051806] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.051806] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore2] e846e43d-68ec-4de3-ba62-e538643b6e4b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.051806] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d852519c-fecc-4f39-8c32-8e3ebdcb58fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.058184] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 830.058184] env[62558]: value = "task-1266743" [ 830.058184] env[62558]: _type = "Task" [ 830.058184] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.068418] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266743, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.080466] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266739, 'name': ReconfigVM_Task, 'duration_secs': 0.61602} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.080760] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 4526df2a-f801-4de8-8218-497a7d22034f/4526df2a-f801-4de8-8218-497a7d22034f.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.081773] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d6399be-e68a-4320-956c-e575cd3f9127 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.088987] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 830.088987] env[62558]: value = "task-1266744" [ 830.088987] env[62558]: _type = "Task" [ 830.088987] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.101292] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266744, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.115721] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0a73ca-b598-4cef-bfb9-81c45ad34a29 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.128113] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Suspending the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 830.128462] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-bc2c2200-db2d-49dd-8ce0-a03170e30016 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.136177] env[62558]: DEBUG oslo_vmware.api [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 830.136177] env[62558]: value = "task-1266745" [ 830.136177] env[62558]: _type = "Task" [ 830.136177] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.150401] env[62558]: DEBUG oslo_vmware.api [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266745, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.159440] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.253602] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.388814] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266741, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.567520] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266743, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.597878] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266744, 'name': Rename_Task, 'duration_secs': 0.193232} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.599384] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.600187] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3591611a-7a50-415e-a0c1-abc5ea285faa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.602912] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c75c3969-6266-491b-9acc-84aba704c5d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.609271] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b82a067-c94a-4065-a3ba-982254c1a32b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.614082] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 830.614082] env[62558]: value = "task-1266746" [ 830.614082] env[62558]: _type = "Task" [ 830.614082] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.646059] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72142f6c-ed24-4968-9ed4-0173fdf4d5d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.652255] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266746, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.660441] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eac8890-ae50-4a1e-96c7-b238b664a529 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.664661] env[62558]: DEBUG oslo_vmware.api [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266745, 'name': SuspendVM_Task} progress is 45%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.679325] env[62558]: DEBUG nova.compute.provider_tree [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.778900] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.889218] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266741, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.73667} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.889530] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.889753] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.890045] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b44acb2-78a3-4434-8c26-c200f8321863 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.898317] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 830.898317] env[62558]: value = "task-1266747" [ 830.898317] env[62558]: _type = "Task" [ 830.898317] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.907640] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.068321] env[62558]: DEBUG oslo_vmware.api [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266743, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.748534} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.068610] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.068741] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.069290] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.069290] env[62558]: INFO nova.compute.manager [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Took 1.62 seconds to destroy the instance on the hypervisor. [ 831.069499] env[62558]: DEBUG oslo.service.loopingcall [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.069717] env[62558]: DEBUG nova.compute.manager [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 831.069810] env[62558]: DEBUG nova.network.neutron [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 831.126315] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266746, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.151853] env[62558]: DEBUG oslo_vmware.api [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266745, 'name': SuspendVM_Task, 'duration_secs': 1.014836} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.152227] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Suspended the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 831.152418] env[62558]: DEBUG nova.compute.manager [None req-050e7264-f806-4936-9794-0af8725a25bb tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.153553] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9aca1a-281c-415d-a97f-2ff450e6b742 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.172073] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.182533] env[62558]: DEBUG nova.scheduler.client.report [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.208621] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.208912] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.209101] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.209318] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.209505] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.209670] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.210034] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.210221] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.210401] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.210575] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.210778] env[62558]: DEBUG nova.virt.hardware [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.211988] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279f81a6-8485-43b4-9b6d-b080e0eb9ca7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.220895] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6045634e-5cdb-40c8-8576-65cdc098cc60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.369119] env[62558]: DEBUG nova.compute.manager [req-78e26cc6-037b-4c63-bcfb-f513e38f78b7 req-209718dd-10a7-4512-9518-0415799ebc96 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Received event network-vif-deleted-8e3c1064-af85-4001-9ba5-1ebba633bbd8 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.369352] env[62558]: INFO nova.compute.manager [req-78e26cc6-037b-4c63-bcfb-f513e38f78b7 req-209718dd-10a7-4512-9518-0415799ebc96 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Neutron deleted interface 8e3c1064-af85-4001-9ba5-1ebba633bbd8; detaching it from the instance and deleting it from the info cache [ 831.369529] env[62558]: DEBUG nova.network.neutron [req-78e26cc6-037b-4c63-bcfb-f513e38f78b7 req-209718dd-10a7-4512-9518-0415799ebc96 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.413020] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1108} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.413020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.413020] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b96f4d-61ec-40c6-bc58-5ae7be10d823 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.433972] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.434461] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c9cb79e-c157-4ec1-ac3e-f84c619f03e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.455305] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 831.455305] env[62558]: value = "task-1266748" [ 831.455305] env[62558]: _type = "Task" [ 831.455305] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.466555] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266748, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.586394] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Successfully updated port: 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.626782] env[62558]: DEBUG oslo_vmware.api [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266746, 'name': PowerOnVM_Task, 'duration_secs': 0.665299} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.627285] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 831.628241] env[62558]: INFO nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Took 9.92 seconds to spawn the instance on the hypervisor. [ 831.628437] env[62558]: DEBUG nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.629297] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351bd74e-d40d-4438-8bf6-56592e0df552 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.688316] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.688844] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.691415] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.351s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.691652] env[62558]: DEBUG nova.objects.instance [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'resources' on Instance uuid 6e01c855-ef25-45c9-a73e-5656e8e05cd4 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.843682] env[62558]: DEBUG nova.network.neutron [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.872833] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c2f102f-3a72-4325-acb8-797ec8de7374 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.882428] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532c9524-eef3-4226-bcd5-65886d8eb865 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.912287] env[62558]: DEBUG nova.compute.manager [req-78e26cc6-037b-4c63-bcfb-f513e38f78b7 req-209718dd-10a7-4512-9518-0415799ebc96 service nova] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Detach interface failed, port_id=8e3c1064-af85-4001-9ba5-1ebba633bbd8, reason: Instance e846e43d-68ec-4de3-ba62-e538643b6e4b could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 831.966070] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266748, 'name': ReconfigVM_Task, 'duration_secs': 0.347056} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.966385] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.967072] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57d6d39e-98b6-4f38-856f-0ecd93d8ddcc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.975038] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 831.975038] env[62558]: value = "task-1266749" [ 831.975038] env[62558]: _type = "Task" [ 831.975038] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.983956] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266749, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.095517] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.095676] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.095871] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.149431] env[62558]: INFO nova.compute.manager [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Took 28.73 seconds to build instance. [ 832.194762] env[62558]: DEBUG nova.compute.utils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.200240] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 832.348638] env[62558]: INFO nova.compute.manager [-] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Took 1.28 seconds to deallocate network for instance. [ 832.492773] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266749, 'name': Rename_Task, 'duration_secs': 0.314599} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.493268] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.493661] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0b0f7b1-b033-43fb-a10d-f10161dbe01d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.505327] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 832.505327] env[62558]: value = "task-1266750" [ 832.505327] env[62558]: _type = "Task" [ 832.505327] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.520550] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266750, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.580201] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abb7897-9108-4b3d-b20e-9005734fb8ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.591384] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40b9009-d014-4128-ab4f-c64dc775ac8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.643873] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6357bf2e-0943-4929-9c1d-eb1baad960f4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.654416] env[62558]: DEBUG oslo_concurrency.lockutils [None req-66221322-11ea-490c-bcad-51d28d183d8e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.562s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.660034] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368913aa-84ba-45f4-9b20-5d4e67f0260c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.691297] env[62558]: DEBUG nova.compute.provider_tree [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.693741] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.702901] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.857011] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.874508] env[62558]: DEBUG nova.network.neutron [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updating instance_info_cache with network_info: [{"id": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "address": "fa:16:3e:20:10:d9", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eb4381-71", "ovs_interfaceid": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.016385] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266750, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.167874] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.200329] env[62558]: DEBUG nova.scheduler.client.report [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.380016] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.380016] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance network_info: |[{"id": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "address": "fa:16:3e:20:10:d9", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eb4381-71", "ovs_interfaceid": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.380203] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:10:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f78b07ea-f425-4622-84f4-706a5d8820a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.386940] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating folder: Project (9a090a1a385e43dc840b63eabb74d0cc). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.388106] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3bae92be-87d2-4eca-8d32-c496481c501c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.396355] env[62558]: DEBUG nova.compute.manager [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received event network-vif-plugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.396678] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.396831] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.397024] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.397257] env[62558]: DEBUG nova.compute.manager [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] No waiting events found dispatching network-vif-plugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.397433] env[62558]: WARNING nova.compute.manager [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received unexpected event network-vif-plugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c for instance with vm_state building and task_state spawning. [ 833.397593] env[62558]: DEBUG nova.compute.manager [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received event network-changed-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.397746] env[62558]: DEBUG nova.compute.manager [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Refreshing instance network info cache due to event network-changed-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.397932] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Acquiring lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.398079] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Acquired lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.398239] env[62558]: DEBUG nova.network.neutron [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Refreshing network info cache for port 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.402591] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created folder: Project (9a090a1a385e43dc840b63eabb74d0cc) in parent group-v272451. [ 833.402591] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating folder: Instances. Parent ref: group-v272523. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.402591] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b36815e-d03b-40c4-922c-6ded69522162 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.412572] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created folder: Instances in parent group-v272523. [ 833.412822] env[62558]: DEBUG oslo.service.loopingcall [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.413127] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.413253] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c863c2b2-8f17-4a5f-baf0-fc9ba91f1875 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.432978] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.432978] env[62558]: value = "task-1266753" [ 833.432978] env[62558]: _type = "Task" [ 833.432978] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.445055] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266753, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.517030] env[62558]: DEBUG oslo_vmware.api [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266750, 'name': PowerOnVM_Task, 'duration_secs': 0.811985} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.517339] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.517545] env[62558]: INFO nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Took 8.18 seconds to spawn the instance on the hypervisor. [ 833.517781] env[62558]: DEBUG nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.518729] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3972957-3a0b-462c-a87a-29082ceda41a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.679270] env[62558]: DEBUG nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.681076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7dcf45-1096-4253-b14e-427565e0dbe5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.694362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.716770] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.025s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.720625] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.723046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.776s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.723303] env[62558]: DEBUG nova.objects.instance [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lazy-loading 'resources' on Instance uuid a8fef137-a651-4724-8ad1-a3e6ddcb5b03 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.750376] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.750658] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.750814] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.750993] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.751374] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.751536] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.751744] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.751901] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.752078] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.752242] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.752412] env[62558]: DEBUG nova.virt.hardware [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.753608] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203b73fa-1755-4e29-86a4-d0aa32b3a1f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.757552] env[62558]: INFO nova.scheduler.client.report [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted allocations for instance 6e01c855-ef25-45c9-a73e-5656e8e05cd4 [ 833.764450] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2115ef77-f277-4aa0-814f-612ee5047221 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.781560] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.787019] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Creating folder: Project (9b7fa7440e864eb2bbd30b77b96d7080). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.787804] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6394775d-ec00-4dd6-b758-dd6f2d1eab07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.800101] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Created folder: Project (9b7fa7440e864eb2bbd30b77b96d7080) in parent group-v272451. [ 833.800663] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Creating folder: Instances. Parent ref: group-v272526. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.800663] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cdde825f-516f-49bd-a588-4c7b10ae9f54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.812376] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Created folder: Instances in parent group-v272526. [ 833.812680] env[62558]: DEBUG oslo.service.loopingcall [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.812951] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.814034] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0723fb23-a272-4e08-98df-425a9b5dfc10 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.834227] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.834227] env[62558]: value = "task-1266756" [ 833.834227] env[62558]: _type = "Task" [ 833.834227] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.842924] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266756, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.946213] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266753, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.038754] env[62558]: INFO nova.compute.manager [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Took 27.71 seconds to build instance. [ 834.166281] env[62558]: DEBUG nova.network.neutron [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updated VIF entry in instance network info cache for port 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.166734] env[62558]: DEBUG nova.network.neutron [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updating instance_info_cache with network_info: [{"id": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "address": "fa:16:3e:20:10:d9", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eb4381-71", "ovs_interfaceid": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.192079] env[62558]: INFO nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] instance snapshotting [ 834.192079] env[62558]: WARNING nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 834.194473] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f35d7c-ce96-4141-88a1-935d0c595e5e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.215619] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9893c95b-630c-46dc-90f0-e643ec4c6124 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.271266] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28d670a5-2533-4d77-a72b-7ebcd46773b6 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "6e01c855-ef25-45c9-a73e-5656e8e05cd4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.972s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.347184] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266756, 'name': CreateVM_Task, 'duration_secs': 0.426345} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.347756] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.348315] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.348528] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.349310] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.350379] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a953cbfe-aaba-4dca-8a8b-b4275c0b915a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.356954] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 834.356954] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ca534-f03f-fe50-81ec-66b425335241" [ 834.356954] env[62558]: _type = "Task" [ 834.356954] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.370979] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ca534-f03f-fe50-81ec-66b425335241, 'name': SearchDatastore_Task, 'duration_secs': 0.01048} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.372053] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.372053] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.372053] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.372053] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.372268] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.372439] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c6c7e81-498f-4049-96fe-f6425c1c2e53 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.383746] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.383957] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.385034] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e380326-63cf-4aa4-b830-ce340703d74e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.394964] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 834.394964] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d139ea-89ec-39a6-fc18-7590fcc8a5a8" [ 834.394964] env[62558]: _type = "Task" [ 834.394964] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.402801] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d139ea-89ec-39a6-fc18-7590fcc8a5a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.446688] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266753, 'name': CreateVM_Task, 'duration_secs': 0.561579} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.447132] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.447846] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.448019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.448779] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.448779] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a627d3cb-dd02-4902-8417-16125afe18fc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.454045] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 834.454045] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52299db7-86c1-0e06-5c28-5306316e3dbd" [ 834.454045] env[62558]: _type = "Task" [ 834.454045] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.463861] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52299db7-86c1-0e06-5c28-5306316e3dbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.542753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-55a768c2-df78-4c61-b10f-3c5d153c8b3b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.395s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.611870] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c68a89-0b89-4ee6-82ac-5b3626d1cc25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.620937] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3fe32e-1cc3-4dd5-878a-82170e30a425 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.652465] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600cad76-ab4e-4a29-a82b-c46bd19d8116 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.660728] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfe5842-b711-4ceb-862a-201c12b7db14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.674947] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47cbf68-5ed7-43da-b506-f517c8ccb0be req-93f6c530-7cc5-46e4-b55a-8768d7c2a65f service nova] Releasing lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.675590] env[62558]: DEBUG nova.compute.provider_tree [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.729039] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 834.729039] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-75578525-3b62-47cc-afe7-b161832c9fb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.737390] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 834.737390] env[62558]: value = "task-1266757" [ 834.737390] env[62558]: _type = "Task" [ 834.737390] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.746742] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266757, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.908523] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d139ea-89ec-39a6-fc18-7590fcc8a5a8, 'name': SearchDatastore_Task, 'duration_secs': 0.008376} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.909672] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5d714e9-85b4-48f5-9355-ac2653a40852 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.916223] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 834.916223] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e8783-31b2-8cfd-c614-8408a45281db" [ 834.916223] env[62558]: _type = "Task" [ 834.916223] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.925776] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e8783-31b2-8cfd-c614-8408a45281db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.931384] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.931647] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.932763] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.932763] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.932763] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.937437] env[62558]: INFO nova.compute.manager [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Terminating instance [ 834.938697] env[62558]: DEBUG nova.compute.manager [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.938918] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.939866] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f49a37-c8df-44f6-98ea-d20771e653de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.948072] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.948682] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71100a3b-07af-444a-914c-8058200540b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.959906] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 834.959906] env[62558]: value = "task-1266758" [ 834.959906] env[62558]: _type = "Task" [ 834.959906] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.967679] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52299db7-86c1-0e06-5c28-5306316e3dbd, 'name': SearchDatastore_Task, 'duration_secs': 0.009257} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.968506] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.968792] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.969102] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.972680] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.045693] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.181802] env[62558]: DEBUG nova.scheduler.client.report [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.247489] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266757, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.425485] env[62558]: DEBUG nova.compute.manager [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-changed-9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.425942] env[62558]: DEBUG nova.compute.manager [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing instance network info cache due to event network-changed-9d645945-9804-4520-bd40-e7475a0dd042. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.426229] env[62558]: DEBUG oslo_concurrency.lockutils [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.426519] env[62558]: DEBUG oslo_concurrency.lockutils [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.426844] env[62558]: DEBUG nova.network.neutron [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing network info cache for port 9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.437164] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e8783-31b2-8cfd-c614-8408a45281db, 'name': SearchDatastore_Task, 'duration_secs': 0.009841} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.438238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.438666] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 66ff3b25-d49c-4129-8b51-5338e75b09b1/66ff3b25-d49c-4129-8b51-5338e75b09b1.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.439403] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.439544] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.439801] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-648f246a-1e23-4288-9272-aab46acdfc55 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.443591] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ea7044d-f1c2-4fdd-8c57-0e810fc67a8f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.453615] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 835.453615] env[62558]: value = "task-1266759" [ 835.453615] env[62558]: _type = "Task" [ 835.453615] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.455221] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.455425] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.459393] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10c83956-e8d9-4c9d-b453-bcbcf3e2386c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.468394] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.470064] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 835.470064] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527e9f20-abd1-9485-3e11-aff4b0fff504" [ 835.470064] env[62558]: _type = "Task" [ 835.470064] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.476132] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266758, 'name': PowerOffVM_Task, 'duration_secs': 0.265381} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.476762] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.476948] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.477231] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c6bbcbd-46d8-47c2-b29d-7162f650f90e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.482246] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527e9f20-abd1-9485-3e11-aff4b0fff504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.525875] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.526116] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.539162] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.539358] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.539546] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleting the datastore file [datastore2] 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.539843] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b100ce95-fb96-4ea8-b189-7a22b00d2e12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.546919] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for the task: (returnval){ [ 835.546919] env[62558]: value = "task-1266761" [ 835.546919] env[62558]: _type = "Task" [ 835.546919] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.561556] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266761, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.568433] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.688340] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.690761] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.196s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.692527] env[62558]: INFO nova.compute.claims [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.710796] env[62558]: INFO nova.scheduler.client.report [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Deleted allocations for instance a8fef137-a651-4724-8ad1-a3e6ddcb5b03 [ 835.749711] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266757, 'name': CreateSnapshot_Task, 'duration_secs': 0.606851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.750275] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 835.750992] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bf06b2-cf6b-4466-9d30-27b805e1a6e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.969386] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266759, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.985388] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527e9f20-abd1-9485-3e11-aff4b0fff504, 'name': SearchDatastore_Task, 'duration_secs': 0.019437} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.987348] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26e658b8-5385-4905-81dc-4bd655fb54ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.996551] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 835.996551] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c9a208-87b0-623c-d8b9-0e9467017107" [ 835.996551] env[62558]: _type = "Task" [ 835.996551] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.010077] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c9a208-87b0-623c-d8b9-0e9467017107, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.059905] env[62558]: DEBUG oslo_vmware.api [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Task: {'id': task-1266761, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274114} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.060383] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 836.060698] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 836.061029] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.061348] env[62558]: INFO nova.compute.manager [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Took 1.12 seconds to destroy the instance on the hypervisor. [ 836.061741] env[62558]: DEBUG oslo.service.loopingcall [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.062086] env[62558]: DEBUG nova.compute.manager [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 836.062246] env[62558]: DEBUG nova.network.neutron [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 836.220300] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c34e524-b9e9-4aa2-b2a7-cbfc9819a88b tempest-InstanceActionsTestJSON-574605109 tempest-InstanceActionsTestJSON-574605109-project-member] Lock "a8fef137-a651-4724-8ad1-a3e6ddcb5b03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.370s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.274110] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 836.277264] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-18b906c0-0de1-4b3b-9023-e513bdd45623 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.285016] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 836.285016] env[62558]: value = "task-1266762" [ 836.285016] env[62558]: _type = "Task" [ 836.285016] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.291825] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266762, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.301867] env[62558]: DEBUG nova.network.neutron [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updated VIF entry in instance network info cache for port 9d645945-9804-4520-bd40-e7475a0dd042. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.302978] env[62558]: DEBUG nova.network.neutron [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.343547] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 836.344785] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d670e1e8-baf5-4783-9de9-c1b7053ef4a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.352569] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 836.352801] env[62558]: ERROR oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk due to incomplete transfer. [ 836.353109] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-56a1c308-e37b-4bd9-9895-c2fb576a4c90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.360998] env[62558]: DEBUG oslo_vmware.rw_handles [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527bd2ac-2618-5187-c5e5-17744a3db26d/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 836.361256] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Uploaded image d7879362-6b66-4ea0-a511-1ecc21020cf6 to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 836.363608] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 836.363873] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2dac171d-212e-4a3b-b401-b3aacb2d6142 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.369626] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 836.369626] env[62558]: value = "task-1266763" [ 836.369626] env[62558]: _type = "Task" [ 836.369626] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.380231] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266763, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.465810] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595641} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.466136] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 66ff3b25-d49c-4129-8b51-5338e75b09b1/66ff3b25-d49c-4129-8b51-5338e75b09b1.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.466413] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.466719] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aece7d8f-4e22-4ef0-8eff-4f0326066768 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.473346] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 836.473346] env[62558]: value = "task-1266764" [ 836.473346] env[62558]: _type = "Task" [ 836.473346] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.481445] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.508654] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c9a208-87b0-623c-d8b9-0e9467017107, 'name': SearchDatastore_Task, 'duration_secs': 0.035734} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.508926] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.509285] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 96361b52-b939-4409-aaf7-ba5b3bf054e7/96361b52-b939-4409-aaf7-ba5b3bf054e7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.509608] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7345f799-52f0-48bd-86ee-a1e86766d24b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.516248] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 836.516248] env[62558]: value = "task-1266765" [ 836.516248] env[62558]: _type = "Task" [ 836.516248] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.524325] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.796185] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266762, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.806952] env[62558]: DEBUG oslo_concurrency.lockutils [req-64b1ed31-2ae6-4f86-b1f6-a1beb2c9e1ad req-d210e205-4a1e-4c4a-93e3-8f409ede24f4 service nova] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.879325] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266763, 'name': Destroy_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.924630] env[62558]: DEBUG nova.network.neutron [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.992433] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067349} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.992433] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.993287] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae54827f-c009-4ff2-a50c-2fa1775f7476 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.024160] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 66ff3b25-d49c-4129-8b51-5338e75b09b1/66ff3b25-d49c-4129-8b51-5338e75b09b1.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.028629] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c6ce548-7cc7-4129-a969-cce31eb066d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.053996] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266765, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.055420] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 837.055420] env[62558]: value = "task-1266766" [ 837.055420] env[62558]: _type = "Task" [ 837.055420] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.071188] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266766, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.187419] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c5fc92-d03f-4a61-a4db-ed462fe5ccb3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.199594] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d656a38-00bc-4307-9dc2-1290f65d308e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.236334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53f6154-c99a-4b24-af08-90793045e9bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.246799] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20e5bc1-6611-4c13-97ec-ccce9809b55a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.260432] env[62558]: DEBUG nova.compute.provider_tree [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.297148] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266762, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.382295] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266763, 'name': Destroy_Task, 'duration_secs': 0.820778} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.382295] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Destroyed the VM [ 837.382544] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 837.382717] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-95fda938-8164-4849-a913-0eaeddeb1b9b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.389394] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 837.389394] env[62558]: value = "task-1266767" [ 837.389394] env[62558]: _type = "Task" [ 837.389394] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.398031] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.428362] env[62558]: INFO nova.compute.manager [-] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Took 1.37 seconds to deallocate network for instance. [ 837.538414] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754168} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.539222] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 96361b52-b939-4409-aaf7-ba5b3bf054e7/96361b52-b939-4409-aaf7-ba5b3bf054e7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.539467] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.539738] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7984d7da-87cc-4a47-9456-2f995c4d480c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.546792] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 837.546792] env[62558]: value = "task-1266768" [ 837.546792] env[62558]: _type = "Task" [ 837.546792] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.554496] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.563900] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266766, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.611370] env[62558]: DEBUG nova.compute.manager [req-a274caf5-1ff4-4140-a601-bbb22c08c05d req-ac6fba48-c6bd-480a-bc07-a778a94b472a service nova] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Received event network-vif-deleted-99b91852-f90d-4a37-9558-042993918fd5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.763959] env[62558]: DEBUG nova.scheduler.client.report [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.797225] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266762, 'name': CloneVM_Task} progress is 95%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.901068] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.939362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.059535] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062894} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.065402] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.065402] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d054c0-3340-471d-a27d-53d2b670967a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.096318] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 96361b52-b939-4409-aaf7-ba5b3bf054e7/96361b52-b939-4409-aaf7-ba5b3bf054e7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.096707] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266766, 'name': ReconfigVM_Task, 'duration_secs': 0.85678} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.096942] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-542e3b31-6171-4cdc-9d5b-4b505df174f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.111336] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 66ff3b25-d49c-4129-8b51-5338e75b09b1/66ff3b25-d49c-4129-8b51-5338e75b09b1.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.112055] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6959e862-a315-44df-8994-b01fec4ce503 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.118827] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 838.118827] env[62558]: value = "task-1266770" [ 838.118827] env[62558]: _type = "Task" [ 838.118827] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.120161] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 838.120161] env[62558]: value = "task-1266769" [ 838.120161] env[62558]: _type = "Task" [ 838.120161] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.133222] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266769, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.136466] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266770, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.268959] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.269704] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.272974] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.681s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.274492] env[62558]: INFO nova.compute.claims [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.297404] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266762, 'name': CloneVM_Task, 'duration_secs': 1.855269} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.297693] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Created linked-clone VM from snapshot [ 838.298478] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9570f309-4780-4c7b-ae5c-9b8906c4cf6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.308725] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Uploading image 9a0277b5-c34d-4c0a-aafd-25eb200450ea {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 838.383670] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 838.383670] env[62558]: value = "vm-272530" [ 838.383670] env[62558]: _type = "VirtualMachine" [ 838.383670] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 838.384045] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e3edf141-4b1d-40a8-bc31-3822570f981e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.392766] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease: (returnval){ [ 838.392766] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f73eb-cec9-be47-088d-d61047287f4a" [ 838.392766] env[62558]: _type = "HttpNfcLease" [ 838.392766] env[62558]: } obtained for exporting VM: (result){ [ 838.392766] env[62558]: value = "vm-272530" [ 838.392766] env[62558]: _type = "VirtualMachine" [ 838.392766] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 838.394301] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the lease: (returnval){ [ 838.394301] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f73eb-cec9-be47-088d-d61047287f4a" [ 838.394301] env[62558]: _type = "HttpNfcLease" [ 838.394301] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 838.405766] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.407344] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 838.407344] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f73eb-cec9-be47-088d-d61047287f4a" [ 838.407344] env[62558]: _type = "HttpNfcLease" [ 838.407344] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 838.636565] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.640378] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266769, 'name': Rename_Task, 'duration_secs': 0.286508} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.640667] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.640931] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-115e79e9-ba75-4559-88f2-830f75bcd499 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.647367] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 838.647367] env[62558]: value = "task-1266772" [ 838.647367] env[62558]: _type = "Task" [ 838.647367] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.656491] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.780908] env[62558]: DEBUG nova.compute.utils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.782417] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.782586] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.838783] env[62558]: DEBUG nova.policy [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca61aea95c774c68a420557e37871537', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b17f6b3fa8314d16bea6d6498f4e9bb0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.902938] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266767, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.906127] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 838.906127] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f73eb-cec9-be47-088d-d61047287f4a" [ 838.906127] env[62558]: _type = "HttpNfcLease" [ 838.906127] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 838.906457] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 838.906457] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f73eb-cec9-be47-088d-d61047287f4a" [ 838.906457] env[62558]: _type = "HttpNfcLease" [ 838.906457] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 838.907113] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626e886a-268e-4589-b810-0509bdad0b36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.914311] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 838.914504] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 839.004703] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d8f33c5d-c285-409c-9661-beadc2cd5e12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.132605] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266770, 'name': ReconfigVM_Task, 'duration_secs': 0.600495} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.132939] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 96361b52-b939-4409-aaf7-ba5b3bf054e7/96361b52-b939-4409-aaf7-ba5b3bf054e7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.133663] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3fd8dc41-0b15-444e-86c9-fab9e3081798 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.139646] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 839.139646] env[62558]: value = "task-1266773" [ 839.139646] env[62558]: _type = "Task" [ 839.139646] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.146320] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Successfully created port: fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.151031] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266773, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.158900] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266772, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.286978] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.403841] env[62558]: DEBUG oslo_vmware.api [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266767, 'name': RemoveSnapshot_Task, 'duration_secs': 1.809804} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.406646] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 839.407127] env[62558]: INFO nova.compute.manager [None req-94b0ad7e-f2dd-4c00-b17f-d20e95529d9c tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 17.86 seconds to snapshot the instance on the hypervisor. [ 839.652012] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266773, 'name': Rename_Task, 'duration_secs': 0.165163} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.655450] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.658016] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b7fd0ec-98a6-4a29-96d6-702b403860bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.665749] env[62558]: DEBUG oslo_vmware.api [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266772, 'name': PowerOnVM_Task, 'duration_secs': 0.807344} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.667160] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.667445] env[62558]: INFO nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Took 5.95 seconds to spawn the instance on the hypervisor. [ 839.667653] env[62558]: DEBUG nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.668231] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 839.668231] env[62558]: value = "task-1266774" [ 839.668231] env[62558]: _type = "Task" [ 839.668231] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.669132] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42bcb47-ac48-4dff-9da3-bfce1edc89b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.686285] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266774, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.701771] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba05b7e-1978-4bc6-99fb-934233fe2d8a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.709483] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe29243a-4f8f-4638-8da4-b7219d46b520 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.747778] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681c6447-876c-4744-9ec9-32476251d224 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.756333] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3da3b9-b889-4b1d-a776-6642a9b38cfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.771423] env[62558]: DEBUG nova.compute.provider_tree [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.181370] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266774, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.193062] env[62558]: INFO nova.compute.manager [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Took 29.55 seconds to build instance. [ 840.274669] env[62558]: DEBUG nova.scheduler.client.report [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.297501] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.322703] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.323021] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.323205] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.323424] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.323632] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.324166] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.324166] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.324302] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.324397] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.324619] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.325158] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.326073] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55e36a7-f496-4f0d-abc7-da6e7e1a0861 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.337075] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb12d25-a753-4d11-ad11-87337d645913 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.397474] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "8acd06fc-b040-468c-980f-6e17b2343c4a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.397595] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.397884] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.398198] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.398423] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.400840] env[62558]: INFO nova.compute.manager [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Terminating instance [ 840.403014] env[62558]: DEBUG nova.compute.manager [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.403273] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.404167] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecaa708-c4ea-41b6-b246-8a011492aaf7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.412825] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.413236] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c237739f-e897-4beb-bced-9c6dc19eba99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.421861] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 840.421861] env[62558]: value = "task-1266775" [ 840.421861] env[62558]: _type = "Task" [ 840.421861] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.430975] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266775, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.658524] env[62558]: DEBUG nova.compute.manager [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Received event network-vif-plugged-fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.658759] env[62558]: DEBUG oslo_concurrency.lockutils [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] Acquiring lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.659036] env[62558]: DEBUG oslo_concurrency.lockutils [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.660845] env[62558]: DEBUG oslo_concurrency.lockutils [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.661090] env[62558]: DEBUG nova.compute.manager [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] No waiting events found dispatching network-vif-plugged-fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.661330] env[62558]: WARNING nova.compute.manager [req-6ff0af62-be1f-4ef2-9f6b-1a49352341f6 req-f8f73b4c-dcaa-4995-84fe-f30941912247 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Received unexpected event network-vif-plugged-fbdcb890-4930-4b7e-8c5c-12082038f0cf for instance with vm_state building and task_state spawning. [ 840.687413] env[62558]: DEBUG oslo_vmware.api [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266774, 'name': PowerOnVM_Task, 'duration_secs': 0.764055} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.688243] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.688243] env[62558]: INFO nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Took 9.52 seconds to spawn the instance on the hypervisor. [ 840.688243] env[62558]: DEBUG nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.689206] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27ab7c9-5248-4c97-a5f1-db155c4aab85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.696516] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1d2d5544-6ac5-4f9d-adb9-74c6a99ca0f9 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.826s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.728745] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Successfully updated port: fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.779912] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.780523] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.783217] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.596s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.783442] env[62558]: DEBUG nova.objects.instance [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 840.935654] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266775, 'name': PowerOffVM_Task, 'duration_secs': 0.359573} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.935967] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.936188] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.936496] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffeda580-65ad-412a-8908-43259ebe49ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.005658] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.005993] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.006327] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Deleting the datastore file [datastore1] 8acd06fc-b040-468c-980f-6e17b2343c4a {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.007287] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff69170d-5c9f-49b3-94fc-f41bef2919ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.015958] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for the task: (returnval){ [ 841.015958] env[62558]: value = "task-1266777" [ 841.015958] env[62558]: _type = "Task" [ 841.015958] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.030832] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.206405] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.212131] env[62558]: INFO nova.compute.manager [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Took 32.73 seconds to build instance. [ 841.231769] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.231962] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.232176] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.288713] env[62558]: DEBUG nova.compute.utils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.293227] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.293406] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.342068] env[62558]: DEBUG nova.policy [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca61aea95c774c68a420557e37871537', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b17f6b3fa8314d16bea6d6498f4e9bb0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.431630] env[62558]: DEBUG nova.compute.manager [None req-3162b27e-3044-4761-9cf1-1cf46a3d221a tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.432687] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7e34cb-9a78-4083-93c9-afd5e1340e6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.508414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.509791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.509791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "66ff3b25-d49c-4129-8b51-5338e75b09b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.509791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.509791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.512876] env[62558]: INFO nova.compute.manager [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Terminating instance [ 841.514922] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "refresh_cache-66ff3b25-d49c-4129-8b51-5338e75b09b1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.515091] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquired lock "refresh_cache-66ff3b25-d49c-4129-8b51-5338e75b09b1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.515263] env[62558]: DEBUG nova.network.neutron [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.525658] env[62558]: DEBUG oslo_vmware.api [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Task: {'id': task-1266777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193066} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.526478] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.526675] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.526856] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.527045] env[62558]: INFO nova.compute.manager [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 841.527293] env[62558]: DEBUG oslo.service.loopingcall [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.527637] env[62558]: DEBUG nova.compute.manager [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.527728] env[62558]: DEBUG nova.network.neutron [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.699871] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Successfully created port: 74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.716079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05625652-2537-464f-9dca-681123fbf582 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.104s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.731029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.785100] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.794284] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.800912] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c7765922-44ec-4892-8888-2dff38578858 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.808021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.156s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.808021] env[62558]: DEBUG nova.objects.instance [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lazy-loading 'resources' on Instance uuid ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.946659] env[62558]: INFO nova.compute.manager [None req-3162b27e-3044-4761-9cf1-1cf46a3d221a tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] instance snapshotting [ 841.947378] env[62558]: DEBUG nova.objects.instance [None req-3162b27e-3044-4761-9cf1-1cf46a3d221a tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lazy-loading 'flavor' on Instance uuid 66ff3b25-d49c-4129-8b51-5338e75b09b1 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.999526] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Updating instance_info_cache with network_info: [{"id": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "address": "fa:16:3e:c5:b3:e2", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbdcb890-49", "ovs_interfaceid": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.034814] env[62558]: DEBUG nova.network.neutron [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.088145] env[62558]: DEBUG nova.network.neutron [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.456936] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1f83f2-7610-469b-877a-57e43ccb870e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.462349] env[62558]: DEBUG nova.network.neutron [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.481671] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9583426-9b7c-47e9-bcfb-c737de93660c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.502487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.502834] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Instance network_info: |[{"id": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "address": "fa:16:3e:c5:b3:e2", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbdcb890-49", "ovs_interfaceid": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.503611] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:b3:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbdcb890-4930-4b7e-8c5c-12082038f0cf', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.511650] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Creating folder: Project (b17f6b3fa8314d16bea6d6498f4e9bb0). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.514442] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62633f4f-45f1-419e-8944-42c645482f5e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.524698] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Created folder: Project (b17f6b3fa8314d16bea6d6498f4e9bb0) in parent group-v272451. [ 842.524911] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Creating folder: Instances. Parent ref: group-v272531. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.525173] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21b9f74e-c66c-4ae7-89ed-000b8420de4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.534657] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Created folder: Instances in parent group-v272531. [ 842.534907] env[62558]: DEBUG oslo.service.loopingcall [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.535124] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.535594] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8d81041-bbca-4653-b26a-97254f53e39a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.556035] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.556035] env[62558]: value = "task-1266780" [ 842.556035] env[62558]: _type = "Task" [ 842.556035] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.563600] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.590460] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Releasing lock "refresh_cache-66ff3b25-d49c-4129-8b51-5338e75b09b1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.590943] env[62558]: DEBUG nova.compute.manager [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.591391] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.592301] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d8860c-f9e8-412b-b7ba-2b4d61072d79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.600444] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.600764] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67ec49a8-2984-46ea-967f-2e6cb30f3b25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.609812] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 842.609812] env[62558]: value = "task-1266781" [ 842.609812] env[62558]: _type = "Task" [ 842.609812] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.619301] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.689873] env[62558]: DEBUG nova.compute.manager [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Received event network-changed-fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.690748] env[62558]: DEBUG nova.compute.manager [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Refreshing instance network info cache due to event network-changed-fbdcb890-4930-4b7e-8c5c-12082038f0cf. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.690748] env[62558]: DEBUG oslo_concurrency.lockutils [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] Acquiring lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.690748] env[62558]: DEBUG oslo_concurrency.lockutils [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] Acquired lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.690748] env[62558]: DEBUG nova.network.neutron [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Refreshing network info cache for port fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.693636] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0059d86-d179-48dd-baba-b02af69c1a23 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.702171] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21fb285-053d-47eb-b352-567ad86ea752 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.734548] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beed5673-d66d-4d56-9d27-bd0e1f3e1ee2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.742516] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ab3bdf-a8e8-49d5-a430-e64588662ef3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.757199] env[62558]: DEBUG nova.compute.provider_tree [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.818082] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.821190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.821190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.821190] env[62558]: INFO nova.compute.manager [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Shelving [ 842.838572] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.838853] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.839035] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.839261] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.839431] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.839593] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.839811] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.839984] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.840189] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.840390] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.840585] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.841736] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2159dbb6-c9ac-4400-8d4f-548c59cf0a63 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.850452] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925f06bd-d932-453b-a43a-cf1368b79887 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.965282] env[62558]: INFO nova.compute.manager [-] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Took 1.44 seconds to deallocate network for instance. [ 842.993349] env[62558]: DEBUG nova.compute.manager [None req-3162b27e-3044-4761-9cf1-1cf46a3d221a tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance disappeared during snapshot {{(pid=62558) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 843.067187] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.119508] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266781, 'name': PowerOffVM_Task, 'duration_secs': 0.16074} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.119766] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.119949] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.120274] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51ed88bf-5bd5-4980-a185-da7e51220e80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.126111] env[62558]: DEBUG nova.compute.manager [None req-3162b27e-3044-4761-9cf1-1cf46a3d221a tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Found 0 images (rotation: 2) {{(pid=62558) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 843.145807] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.146110] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.146958] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Deleting the datastore file [datastore1] 66ff3b25-d49c-4129-8b51-5338e75b09b1 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.146958] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55252133-16ef-40d9-8b5f-5b2362741390 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.153732] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for the task: (returnval){ [ 843.153732] env[62558]: value = "task-1266783" [ 843.153732] env[62558]: _type = "Task" [ 843.153732] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.163050] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266783, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.260121] env[62558]: DEBUG nova.scheduler.client.report [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.328528] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 843.328826] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc6c0855-4fd7-4c2f-8d60-fe0fcf88d2fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.334876] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 843.334876] env[62558]: value = "task-1266784" [ 843.334876] env[62558]: _type = "Task" [ 843.334876] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.343962] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.352562] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Successfully updated port: 74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.450032] env[62558]: DEBUG nova.network.neutron [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Updated VIF entry in instance network info cache for port fbdcb890-4930-4b7e-8c5c-12082038f0cf. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.450371] env[62558]: DEBUG nova.network.neutron [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Updating instance_info_cache with network_info: [{"id": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "address": "fa:16:3e:c5:b3:e2", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbdcb890-49", "ovs_interfaceid": "fbdcb890-4930-4b7e-8c5c-12082038f0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.472513] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.568163] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.663378] env[62558]: DEBUG oslo_vmware.api [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Task: {'id': task-1266783, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138971} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.663633] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.663846] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.664053] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.664222] env[62558]: INFO nova.compute.manager [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Took 1.07 seconds to destroy the instance on the hypervisor. [ 843.664510] env[62558]: DEBUG oslo.service.loopingcall [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.664784] env[62558]: DEBUG nova.compute.manager [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 843.664888] env[62558]: DEBUG nova.network.neutron [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 843.680297] env[62558]: DEBUG nova.network.neutron [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.767956] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.770728] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.097s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.771016] env[62558]: DEBUG nova.objects.instance [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lazy-loading 'resources' on Instance uuid ebd84b3f-0c5d-40e3-aa70-f8b3054b7109 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.792645] env[62558]: INFO nova.scheduler.client.report [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleted allocations for instance ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55 [ 843.846868] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266784, 'name': PowerOffVM_Task, 'duration_secs': 0.255714} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.847164] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.847946] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f46648f-cb73-447a-b234-dd67ef57a8b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.866730] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.866903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.867081] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.872014] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd39576b-72f5-40bb-8546-e0d94926598f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.953314] env[62558]: DEBUG oslo_concurrency.lockutils [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] Releasing lock "refresh_cache-6243631f-b240-4d7c-8910-0bf3d2dedb77" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.953785] env[62558]: DEBUG nova.compute.manager [req-7c96e2e1-3ada-4773-bdaa-db1d4e144039 req-2ced4fda-5269-40bb-986b-3996b4b11042 service nova] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Received event network-vif-deleted-23817056-3d15-4a89-8f18-84e4f7549efe {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.067937] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.183964] env[62558]: DEBUG nova.network.neutron [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.301780] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3dfc27fb-a921-4c25-93d2-e5189f33766b tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.816s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.382540] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 844.383403] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-30755a3f-91e0-404c-8da0-b1ec1aeabf2b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.393093] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 844.393093] env[62558]: value = "task-1266785" [ 844.393093] env[62558]: _type = "Task" [ 844.393093] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.404363] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266785, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.412496] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.568176] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.598809] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Updating instance_info_cache with network_info: [{"id": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "address": "fa:16:3e:07:8f:17", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74146c35-c7", "ovs_interfaceid": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.608734] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f69bb4-4b4f-4d37-95f8-f39e93b56b2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.617600] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182e42e6-4c6d-4a8c-a040-13310901230e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.648702] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100294fe-c033-46ae-afee-63d0365b02d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.657233] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0135e9-c121-4d87-a55b-3ec9e435d624 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.672516] env[62558]: DEBUG nova.compute.provider_tree [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.685369] env[62558]: INFO nova.compute.manager [-] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Took 1.02 seconds to deallocate network for instance. [ 844.715675] env[62558]: DEBUG nova.compute.manager [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Received event network-vif-plugged-74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.715958] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Acquiring lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.715958] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.716205] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.716256] env[62558]: DEBUG nova.compute.manager [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] No waiting events found dispatching network-vif-plugged-74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.716410] env[62558]: WARNING nova.compute.manager [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Received unexpected event network-vif-plugged-74146c35-c7d1-4654-9100-87b81d3d4ea4 for instance with vm_state building and task_state spawning. [ 844.716570] env[62558]: DEBUG nova.compute.manager [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Received event network-changed-74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.716721] env[62558]: DEBUG nova.compute.manager [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Refreshing instance network info cache due to event network-changed-74146c35-c7d1-4654-9100-87b81d3d4ea4. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.716893] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Acquiring lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.905178] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266785, 'name': CreateSnapshot_Task, 'duration_secs': 0.457461} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.905484] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 844.906296] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c5a641-f0d2-4799-8266-8d1f2935c0cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.068475] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266780, 'name': CreateVM_Task, 'duration_secs': 2.128574} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.068716] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.069405] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.069582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.069910] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.070207] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-126e7813-41be-47d5-be14-8533706ddede {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.074971] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 845.074971] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5283080c-9bf4-7683-fc24-d63c4f66faac" [ 845.074971] env[62558]: _type = "Task" [ 845.074971] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.084274] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5283080c-9bf4-7683-fc24-d63c4f66faac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.102046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.102361] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Instance network_info: |[{"id": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "address": "fa:16:3e:07:8f:17", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74146c35-c7", "ovs_interfaceid": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.102736] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Acquired lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.102943] env[62558]: DEBUG nova.network.neutron [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Refreshing network info cache for port 74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.104410] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:8f:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74146c35-c7d1-4654-9100-87b81d3d4ea4', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.113365] env[62558]: DEBUG oslo.service.loopingcall [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.114404] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.114649] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1110770d-9a4d-491e-9c5a-1fb23636cf5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.135330] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.135330] env[62558]: value = "task-1266786" [ 845.135330] env[62558]: _type = "Task" [ 845.135330] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.143283] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266786, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.168615] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.168839] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.175472] env[62558]: DEBUG nova.scheduler.client.report [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.191867] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.427064] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 845.427445] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ceeeb8f6-c55e-4b63-af26-0b3934747b3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.436794] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 845.436794] env[62558]: value = "task-1266787" [ 845.436794] env[62558]: _type = "Task" [ 845.436794] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.445422] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266787, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.588113] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5283080c-9bf4-7683-fc24-d63c4f66faac, 'name': SearchDatastore_Task, 'duration_secs': 0.013179} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.588527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.588877] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.589233] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.589401] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.589590] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.589961] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cca895a-6495-43b8-aaef-d8b8b7ca8565 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.598408] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.598596] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.599380] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4befb89f-ee91-48af-b744-ce2fe0448c91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.604719] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 845.604719] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b29c20-7552-0082-6291-c65a83282f92" [ 845.604719] env[62558]: _type = "Task" [ 845.604719] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.612700] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b29c20-7552-0082-6291-c65a83282f92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.646434] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266786, 'name': CreateVM_Task, 'duration_secs': 0.45112} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.646774] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.648083] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.648083] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.648083] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.648278] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa7df034-a9a7-4bc4-91d7-68f70f285610 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.653388] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 845.653388] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a4f813-704b-de0a-47a6-6e1a361d8c9f" [ 845.653388] env[62558]: _type = "Task" [ 845.653388] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.663488] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a4f813-704b-de0a-47a6-6e1a361d8c9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.675572] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 845.675757] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 845.675878] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 845.679965] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.684133] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.702s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.684133] env[62558]: INFO nova.compute.claims [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.705251] env[62558]: INFO nova.scheduler.client.report [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Deleted allocations for instance ebd84b3f-0c5d-40e3-aa70-f8b3054b7109 [ 845.895353] env[62558]: DEBUG nova.network.neutron [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Updated VIF entry in instance network info cache for port 74146c35-c7d1-4654-9100-87b81d3d4ea4. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.895556] env[62558]: DEBUG nova.network.neutron [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Updating instance_info_cache with network_info: [{"id": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "address": "fa:16:3e:07:8f:17", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74146c35-c7", "ovs_interfaceid": "74146c35-c7d1-4654-9100-87b81d3d4ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.950014] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266787, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.116369] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b29c20-7552-0082-6291-c65a83282f92, 'name': SearchDatastore_Task, 'duration_secs': 0.013217} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.117544] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-979700a7-fd1a-49d3-9f5f-d77590f44f27 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.122574] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 846.122574] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a38503-41bb-17f5-46e3-15714ba43bfa" [ 846.122574] env[62558]: _type = "Task" [ 846.122574] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.131477] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a38503-41bb-17f5-46e3-15714ba43bfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.164231] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a4f813-704b-de0a-47a6-6e1a361d8c9f, 'name': SearchDatastore_Task, 'duration_secs': 0.022541} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.164609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.164861] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.165096] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.182000] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 846.182328] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 846.192901] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 846.193797] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba5e06da-2aba-4466-bc14-fcfad57cd415 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.200964] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 846.200964] env[62558]: ERROR oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk due to incomplete transfer. [ 846.201614] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d6764c0b-f928-4b33-8a6f-046cd4d0b810 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.206630] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.206768] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.206916] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 846.207082] env[62558]: DEBUG nova.objects.instance [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lazy-loading 'info_cache' on Instance uuid 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.209121] env[62558]: DEBUG oslo_vmware.rw_handles [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52462082-3259-1105-13aa-3edbed8b1dc0/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 846.209304] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Uploaded image 9a0277b5-c34d-4c0a-aafd-25eb200450ea to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 846.211568] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 846.214028] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-27a73598-d668-4d44-8a06-e70f58098953 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.216136] env[62558]: DEBUG oslo_concurrency.lockutils [None req-edb49b0e-8b75-4274-a073-cc178bacefa7 tempest-MultipleCreateTestJSON-1883034578 tempest-MultipleCreateTestJSON-1883034578-project-member] Lock "ebd84b3f-0c5d-40e3-aa70-f8b3054b7109" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.653s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.221353] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 846.221353] env[62558]: value = "task-1266788" [ 846.221353] env[62558]: _type = "Task" [ 846.221353] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.229857] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266788, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.398343] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd1f703b-e39b-4f74-aa74-11bbd69b2a16 req-f1d6b114-3b64-4ef9-a21d-001b62749d61 service nova] Releasing lock "refresh_cache-2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.448364] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266787, 'name': CloneVM_Task} progress is 95%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.634028] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a38503-41bb-17f5-46e3-15714ba43bfa, 'name': SearchDatastore_Task, 'duration_secs': 0.009395} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.634028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.634028] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6243631f-b240-4d7c-8910-0bf3d2dedb77/6243631f-b240-4d7c-8910-0bf3d2dedb77.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.634028] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.634464] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.634464] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1782f59d-b031-4dd5-9d05-30b09d8c79f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.636358] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ce9b813-d0a9-49b3-943a-d23817c430d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.643702] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 846.643702] env[62558]: value = "task-1266789" [ 846.643702] env[62558]: _type = "Task" [ 846.643702] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.648353] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.648595] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.649710] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f7dfe5a-4f97-47d7-8eb8-373f9f4bc56c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.654494] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266789, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.657225] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 846.657225] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5242e4f6-6ed4-4001-36bb-3cd7bdffc875" [ 846.657225] env[62558]: _type = "Task" [ 846.657225] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.664273] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5242e4f6-6ed4-4001-36bb-3cd7bdffc875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.733867] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266788, 'name': Destroy_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.954281] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266787, 'name': CloneVM_Task, 'duration_secs': 1.355505} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.957925] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Created linked-clone VM from snapshot [ 846.959209] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febe0148-9f96-48e5-83a0-a33b66d268b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.968940] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Uploading image 2d406018-3303-46ca-b943-5ea7a0ad206c {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 847.000606] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 847.000606] env[62558]: value = "vm-272536" [ 847.000606] env[62558]: _type = "VirtualMachine" [ 847.000606] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 847.000957] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ac323f24-286c-4c48-85d0-9114ec51b0a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.013022] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lease: (returnval){ [ 847.013022] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eb8a41-1d65-6d4b-0ba4-74c0187abd5b" [ 847.013022] env[62558]: _type = "HttpNfcLease" [ 847.013022] env[62558]: } obtained for exporting VM: (result){ [ 847.013022] env[62558]: value = "vm-272536" [ 847.013022] env[62558]: _type = "VirtualMachine" [ 847.013022] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 847.013022] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the lease: (returnval){ [ 847.013022] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eb8a41-1d65-6d4b-0ba4-74c0187abd5b" [ 847.013022] env[62558]: _type = "HttpNfcLease" [ 847.013022] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 847.020408] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35b3d21-828b-4889-b9ef-d59402700510 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.025821] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.025821] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eb8a41-1d65-6d4b-0ba4-74c0187abd5b" [ 847.025821] env[62558]: _type = "HttpNfcLease" [ 847.025821] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 847.032921] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be6ade5-07ef-4f17-bb15-ea790cc10ed0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.071624] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175a7b5c-1185-43f1-8af5-750e2951e421 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.082064] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e02920f-1b6f-4b18-bbf3-59a1afb6fc37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.102048] env[62558]: DEBUG nova.compute.provider_tree [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.153315] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266789, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498143} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.153904] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 6243631f-b240-4d7c-8910-0bf3d2dedb77/6243631f-b240-4d7c-8910-0bf3d2dedb77.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.154974] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.154974] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-956d5796-61f5-4968-847a-9bdf2075cab5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.162090] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 847.162090] env[62558]: value = "task-1266791" [ 847.162090] env[62558]: _type = "Task" [ 847.162090] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.170595] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5242e4f6-6ed4-4001-36bb-3cd7bdffc875, 'name': SearchDatastore_Task, 'duration_secs': 0.021575} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.171929] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3832d14-f614-4adf-8907-2c5b44cd5b51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.177142] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266791, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.180139] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 847.180139] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526afb57-39b1-c88c-ec98-5adeb17147cc" [ 847.180139] env[62558]: _type = "Task" [ 847.180139] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.188749] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526afb57-39b1-c88c-ec98-5adeb17147cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.234108] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266788, 'name': Destroy_Task, 'duration_secs': 0.840189} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.234859] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.236303] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Destroyed the VM [ 847.236671] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 847.237242] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-53cf503b-2ad0-4786-a3fc-678999885fde {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.255462] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 847.255462] env[62558]: value = "task-1266792" [ 847.255462] env[62558]: _type = "Task" [ 847.255462] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.265697] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266792, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.399062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.399360] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.522756] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.522756] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eb8a41-1d65-6d4b-0ba4-74c0187abd5b" [ 847.522756] env[62558]: _type = "HttpNfcLease" [ 847.522756] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 847.523116] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 847.523116] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eb8a41-1d65-6d4b-0ba4-74c0187abd5b" [ 847.523116] env[62558]: _type = "HttpNfcLease" [ 847.523116] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 847.527041] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afee31b-edf4-4b5f-9f7c-c7c39c52dbab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.531904] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 847.532383] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 847.604736] env[62558]: DEBUG nova.scheduler.client.report [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.649038] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c829346b-e651-4b4d-ac6d-03f1124e806e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.674292] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266791, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098101} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.677893] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.678923] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9570faad-0466-439c-a67e-6d9d2affb388 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.710703] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 6243631f-b240-4d7c-8910-0bf3d2dedb77/6243631f-b240-4d7c-8910-0bf3d2dedb77.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.719285] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff36b62b-7cc2-4541-8956-4865c3473961 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.730559] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526afb57-39b1-c88c-ec98-5adeb17147cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009914} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.730891] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.732147] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca/2a7b4e08-d9b1-49f1-9611-ca1acc3315ca.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.732147] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-555c6e9e-bca7-47a1-b8d9-1ac97aeed836 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.736472] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 847.736472] env[62558]: value = "task-1266793" [ 847.736472] env[62558]: _type = "Task" [ 847.736472] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.743111] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 847.743111] env[62558]: value = "task-1266794" [ 847.743111] env[62558]: _type = "Task" [ 847.743111] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.749991] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.756341] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.764951] env[62558]: DEBUG oslo_vmware.api [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266792, 'name': RemoveSnapshot_Task, 'duration_secs': 0.456659} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.765142] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 847.765797] env[62558]: INFO nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Took 13.57 seconds to snapshot the instance on the hypervisor. [ 847.896451] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.897010] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.899467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.899467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.899467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.900162] env[62558]: INFO nova.compute.manager [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Terminating instance [ 847.903059] env[62558]: DEBUG nova.compute.utils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.904273] env[62558]: DEBUG nova.compute.manager [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.904739] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.905895] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aec8d75-4822-4aa2-a6b9-26dc68a09e87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.915687] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.917148] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e45ac1f-ef4e-4ba9-b44b-abf50a59f9f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.928092] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.013688] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.014103] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.014477] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleting the datastore file [datastore2] 4a3f7642-1b9c-40d8-973b-5153b559bda0 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.014939] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d63bd12f-cd19-4e8a-b131-59cbf3363f93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.023945] env[62558]: DEBUG oslo_vmware.api [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 848.023945] env[62558]: value = "task-1266796" [ 848.023945] env[62558]: _type = "Task" [ 848.023945] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.036230] env[62558]: DEBUG oslo_vmware.api [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.111838] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.113936] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.116211] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.391s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.117378] env[62558]: DEBUG nova.objects.instance [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'resources' on Instance uuid 6d89107f-0727-4b8c-863d-d1e635000bff {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.255689] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266793, 'name': ReconfigVM_Task, 'duration_secs': 0.327382} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.259709] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 6243631f-b240-4d7c-8910-0bf3d2dedb77/6243631f-b240-4d7c-8910-0bf3d2dedb77.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.261097] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500811} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.261723] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c86595d1-2565-4a34-b2ea-0a8c7fca4458 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.264446] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca/2a7b4e08-d9b1-49f1-9611-ca1acc3315ca.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.264694] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.265558] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc7325a3-a0de-461b-be3e-a5f3641dd645 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.268962] env[62558]: DEBUG nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance disappeared during snapshot {{(pid=62558) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 848.275534] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 848.275534] env[62558]: value = "task-1266797" [ 848.275534] env[62558]: _type = "Task" [ 848.275534] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.281298] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 848.281298] env[62558]: value = "task-1266798" [ 848.281298] env[62558]: _type = "Task" [ 848.281298] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.290045] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266797, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.295861] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.302016] env[62558]: DEBUG nova.compute.manager [None req-c5fe835c-8f78-4bcd-bf16-9e529b853319 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image not found during clean up 9a0277b5-c34d-4c0a-aafd-25eb200450ea {{(pid=62558) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 848.410577] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.430632] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-867c0fb3-d4f6-4945-a5cc-6d93efcadb79" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.431880] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 848.431880] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.431880] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.431880] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.431880] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.431880] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.432321] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.432477] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 848.432635] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.536687] env[62558]: DEBUG oslo_vmware.api [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354454} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.536816] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.537343] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.538186] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.538444] env[62558]: INFO nova.compute.manager [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Took 0.63 seconds to destroy the instance on the hypervisor. [ 848.538888] env[62558]: DEBUG oslo.service.loopingcall [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.538888] env[62558]: DEBUG nova.compute.manager [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.539034] env[62558]: DEBUG nova.network.neutron [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.622504] env[62558]: DEBUG nova.compute.utils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.629895] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.630193] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.714551] env[62558]: DEBUG nova.policy [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca61aea95c774c68a420557e37871537', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b17f6b3fa8314d16bea6d6498f4e9bb0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.793884] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266797, 'name': Rename_Task, 'duration_secs': 0.180858} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.796620] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.797929] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c717dfd-a061-442f-a721-7183ec1c3ceb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.802999] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081996} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.803718] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.804813] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747d9f67-0538-4aa5-97e4-fc8eae3d204f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.809476] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 848.809476] env[62558]: value = "task-1266799" [ 848.809476] env[62558]: _type = "Task" [ 848.809476] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.843378] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca/2a7b4e08-d9b1-49f1-9611-ca1acc3315ca.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.844764] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6871d1c-fe5d-419e-b907-9a9bd7cd4049 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.868794] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266799, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.879877] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 848.879877] env[62558]: value = "task-1266800" [ 848.879877] env[62558]: _type = "Task" [ 848.879877] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.891159] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266800, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.936240] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.990779] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.991112] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.079831] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Successfully created port: 332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.113711] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6175696-cf81-4512-af63-22c4f89c2e86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.122290] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ff8de1-1038-4f08-ab1e-2f0c9c116ace {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.160083] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.166021] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e972153-4504-4c4c-b42e-9e612b5b9067 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.176158] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d59567-58f0-489b-8a5e-d9f9dd1588ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.198849] env[62558]: DEBUG nova.compute.provider_tree [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.300653] env[62558]: DEBUG nova.compute.manager [req-de55ec61-9cba-4fbe-a880-c806d89a4aa8 req-da31d124-9b75-4aed-b46a-6267d69cbff3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Received event network-vif-deleted-8c7dccfe-46a4-49d4-8622-3e385f26e26c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.301047] env[62558]: INFO nova.compute.manager [req-de55ec61-9cba-4fbe-a880-c806d89a4aa8 req-da31d124-9b75-4aed-b46a-6267d69cbff3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Neutron deleted interface 8c7dccfe-46a4-49d4-8622-3e385f26e26c; detaching it from the instance and deleting it from the info cache [ 849.301440] env[62558]: DEBUG nova.network.neutron [req-de55ec61-9cba-4fbe-a880-c806d89a4aa8 req-da31d124-9b75-4aed-b46a-6267d69cbff3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.320590] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266799, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.393092] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266800, 'name': ReconfigVM_Task, 'duration_secs': 0.406171} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.393092] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca/2a7b4e08-d9b1-49f1-9611-ca1acc3315ca.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.393092] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8366f2ec-ecd1-44e1-8920-640609a01de8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.401502] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 849.401502] env[62558]: value = "task-1266801" [ 849.401502] env[62558]: _type = "Task" [ 849.401502] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.411353] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266801, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.432555] env[62558]: DEBUG nova.network.neutron [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.495089] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.550481] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.550481] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.550741] env[62558]: INFO nova.compute.manager [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Attaching volume ac9c0aa2-f602-4f72-a5d5-7e7027472d3c to /dev/sdb [ 849.590030] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6363e369-78cf-4901-9d78-8c50b434c96b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.597391] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f538524e-123d-473e-b281-a6c98633ce81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.613890] env[62558]: DEBUG nova.virt.block_device [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating existing volume attachment record: 1f5d22dd-6c8d-4de9-a8c5-288833ae1fc4 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 849.704074] env[62558]: DEBUG nova.scheduler.client.report [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.805258] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a502a53c-aa8c-4b5e-8884-3ec62869f1ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.816704] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57212d2d-72de-4bf6-896b-99ebb806b69d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.832461] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266799, 'name': PowerOnVM_Task, 'duration_secs': 0.607976} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.832804] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.833039] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Took 9.54 seconds to spawn the instance on the hypervisor. [ 849.833240] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.834028] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e94234-a595-4b81-bbd1-0d7a5e22c377 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.846327] env[62558]: DEBUG nova.compute.manager [req-de55ec61-9cba-4fbe-a880-c806d89a4aa8 req-da31d124-9b75-4aed-b46a-6267d69cbff3 service nova] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Detach interface failed, port_id=8c7dccfe-46a4-49d4-8622-3e385f26e26c, reason: Instance 4a3f7642-1b9c-40d8-973b-5153b559bda0 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 849.910819] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266801, 'name': Rename_Task, 'duration_secs': 0.133421} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.911238] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.911552] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9caf75c-0525-485b-8145-21623ed38f5e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.918339] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 849.918339] env[62558]: value = "task-1266805" [ 849.918339] env[62558]: _type = "Task" [ 849.918339] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.926769] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.935537] env[62558]: INFO nova.compute.manager [-] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Took 1.40 seconds to deallocate network for instance. [ 850.021908] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.172018] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.202046] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.202046] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.202235] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.202328] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.202526] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.202702] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.202927] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.203111] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.203296] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.203468] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.203631] env[62558]: DEBUG nova.virt.hardware [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.204817] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcb66fe-004a-46ff-8fc6-747ab4e3f0ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.209657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.211856] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.772s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.213433] env[62558]: INFO nova.compute.claims [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.222673] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ddabc2-8f63-4814-afa2-f38bd3aebb58 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.239871] env[62558]: INFO nova.scheduler.client.report [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocations for instance 6d89107f-0727-4b8c-863d-d1e635000bff [ 850.362465] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Took 34.89 seconds to build instance. [ 850.428090] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266805, 'name': PowerOnVM_Task, 'duration_secs': 0.48485} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.428384] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.428611] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Took 7.61 seconds to spawn the instance on the hypervisor. [ 850.428825] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.429669] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7df332-306c-48ae-812c-c62e86036cab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.443545] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.553060] env[62558]: DEBUG nova.compute.manager [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Received event network-vif-plugged-332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.553831] env[62558]: DEBUG oslo_concurrency.lockutils [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] Acquiring lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.554379] env[62558]: DEBUG oslo_concurrency.lockutils [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.554643] env[62558]: DEBUG oslo_concurrency.lockutils [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.554927] env[62558]: DEBUG nova.compute.manager [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] No waiting events found dispatching network-vif-plugged-332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.555189] env[62558]: WARNING nova.compute.manager [req-09bd6882-a7a7-4c24-aec2-f09ba48206ce req-98d0f528-8d31-46af-85a7-5b1e247c7383 service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Received unexpected event network-vif-plugged-332b0302-a7b5-4574-b8ef-006d49069a3a for instance with vm_state building and task_state spawning. [ 850.647883] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Successfully updated port: 332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.747537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a00a31d1-0455-4311-ae25-02c4f1d7fb63 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "6d89107f-0727-4b8c-863d-d1e635000bff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.217s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.864778] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.585s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.946539] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Took 35.38 seconds to build instance. [ 851.151865] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.152017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.152158] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.450107] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.139s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.540650] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c163c06b-cd3d-4671-9b0a-75b08e644cdc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.548498] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36864047-fe65-4eeb-a0a9-dd3ffd6d77ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.580861] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412ecbd3-e4e5-4795-a0c1-9b4bc9b777f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.588499] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d56e01-565f-43ba-92cf-0c2838800bfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.602013] env[62558]: DEBUG nova.compute.provider_tree [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.683027] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.866418] env[62558]: DEBUG nova.network.neutron [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Updating instance_info_cache with network_info: [{"id": "332b0302-a7b5-4574-b8ef-006d49069a3a", "address": "fa:16:3e:a7:c3:01", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap332b0302-a7", "ovs_interfaceid": "332b0302-a7b5-4574-b8ef-006d49069a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.105329] env[62558]: DEBUG nova.scheduler.client.report [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.369045] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.369376] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Instance network_info: |[{"id": "332b0302-a7b5-4574-b8ef-006d49069a3a", "address": "fa:16:3e:a7:c3:01", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap332b0302-a7", "ovs_interfaceid": "332b0302-a7b5-4574-b8ef-006d49069a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.369952] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:c3:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55520f67-d092-4eb7-940f-d7cceaa1ca1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '332b0302-a7b5-4574-b8ef-006d49069a3a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.377939] env[62558]: DEBUG oslo.service.loopingcall [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.378212] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.378466] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53149ee1-655b-47fe-b68e-695c17a73c1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.398348] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.398348] env[62558]: value = "task-1266807" [ 852.398348] env[62558]: _type = "Task" [ 852.398348] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.405797] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266807, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.613152] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.613416] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.617501] env[62558]: DEBUG nova.compute.manager [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Received event network-changed-332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.617718] env[62558]: DEBUG nova.compute.manager [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Refreshing instance network info cache due to event network-changed-332b0302-a7b5-4574-b8ef-006d49069a3a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.617978] env[62558]: DEBUG oslo_concurrency.lockutils [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] Acquiring lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.618120] env[62558]: DEBUG oslo_concurrency.lockutils [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] Acquired lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.618290] env[62558]: DEBUG nova.network.neutron [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Refreshing network info cache for port 332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.620118] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.230s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.621200] env[62558]: INFO nova.compute.claims [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.911359] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266807, 'name': CreateVM_Task, 'duration_secs': 0.408588} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.911820] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.912351] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.912580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.912966] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.913290] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33344889-dff2-4ba5-acfc-ae4385f9ac0a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.918666] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 852.918666] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52913c23-47c7-454a-6f3a-fb1a73200d19" [ 852.918666] env[62558]: _type = "Task" [ 852.918666] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.929745] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52913c23-47c7-454a-6f3a-fb1a73200d19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.121362] env[62558]: DEBUG nova.compute.utils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.122998] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.123188] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.168514] env[62558]: DEBUG nova.policy [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62047bfc3c2a42d8b860f618d0215241', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e75f3bcc65ee48d28da7ee016b2fd4f5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.392631] env[62558]: DEBUG nova.network.neutron [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Updated VIF entry in instance network info cache for port 332b0302-a7b5-4574-b8ef-006d49069a3a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.393035] env[62558]: DEBUG nova.network.neutron [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Updating instance_info_cache with network_info: [{"id": "332b0302-a7b5-4574-b8ef-006d49069a3a", "address": "fa:16:3e:a7:c3:01", "network": {"id": "c8b5975e-32cf-441a-987f-fafa3a106e99", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1407998697-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b17f6b3fa8314d16bea6d6498f4e9bb0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55520f67-d092-4eb7-940f-d7cceaa1ca1c", "external-id": "nsx-vlan-transportzone-717", "segmentation_id": 717, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap332b0302-a7", "ovs_interfaceid": "332b0302-a7b5-4574-b8ef-006d49069a3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.431570] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52913c23-47c7-454a-6f3a-fb1a73200d19, 'name': SearchDatastore_Task, 'duration_secs': 0.011451} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.431570] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.431570] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.431570] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.431954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.431954] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.432169] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ac60bc7-74cd-4991-8a3b-a9efe1256029 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.442411] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.442664] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.443438] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-314fbd6f-208a-436f-b796-0778794834c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.446980] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Successfully created port: 1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.450318] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 853.450318] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cc3631-cb4e-6816-2294-d1c92787bd42" [ 853.450318] env[62558]: _type = "Task" [ 853.450318] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.457773] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cc3631-cb4e-6816-2294-d1c92787bd42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.627341] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.895873] env[62558]: DEBUG oslo_concurrency.lockutils [req-53ee9ba6-6ae2-4365-9bc4-c4d8862e6c95 req-ca474b9b-41b4-4108-a169-db61ee07731a service nova] Releasing lock "refresh_cache-a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.963871] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52cc3631-cb4e-6816-2294-d1c92787bd42, 'name': SearchDatastore_Task, 'duration_secs': 0.013102} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.964620] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0db7817a-fb4e-48b1-b912-ba0949907b3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.972164] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 853.972164] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ebd39-a835-0eaf-1323-65e6f058811e" [ 853.972164] env[62558]: _type = "Task" [ 853.972164] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.980684] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ebd39-a835-0eaf-1323-65e6f058811e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.982456] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64249900-c3b1-42c0-93e2-24f8f0f7a9d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.989927] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036bf100-90ea-44fc-9288-83ce4dc301f4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.020109] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f7c064-94e0-4401-be5f-4a7ad1501c95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.028280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b84b311-ec8e-48e6-a5c9-f4f4cadc619d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.043555] env[62558]: DEBUG nova.compute.provider_tree [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.167213] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 854.167470] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272538', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'name': 'volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1d242e6-1561-4bd4-8e39-281ab6346661', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'serial': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 854.168573] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f986b42-642f-45db-ba12-9db2b0ff57c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.186364] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a82effc-5be9-4b38-a688-46342bc9c5f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.211119] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c/volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.211398] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b091ac0a-a083-45aa-9e4e-5951c5cc52fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.230188] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 854.230188] env[62558]: value = "task-1266808" [ 854.230188] env[62558]: _type = "Task" [ 854.230188] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.238298] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266808, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.483011] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ebd39-a835-0eaf-1323-65e6f058811e, 'name': SearchDatastore_Task, 'duration_secs': 0.016219} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.483326] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.483624] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884/a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.483901] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a865f62b-d6ee-4930-9e7f-57a1b77f1497 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.490131] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 854.490131] env[62558]: value = "task-1266809" [ 854.490131] env[62558]: _type = "Task" [ 854.490131] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.498575] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.546867] env[62558]: DEBUG nova.scheduler.client.report [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.636673] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.662408] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.662710] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.662878] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.663111] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.663283] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.663452] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.663667] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.663851] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.664043] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.664212] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.664407] env[62558]: DEBUG nova.virt.hardware [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.665303] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b3e476-6349-46e1-9ac6-b471fd84faf2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.673403] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d517d2-7e1e-4c1a-8b25-3d38e28182a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.741074] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.768259] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 854.769163] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4c895a-a0a7-459f-8ee3-f70ca39b865c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.775383] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 854.775551] env[62558]: ERROR oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk due to incomplete transfer. [ 854.775773] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1b32d013-4bd6-46bc-a3c5-7906798f72c8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.783240] env[62558]: DEBUG oslo_vmware.rw_handles [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5294b17b-3d7b-c86b-595a-ec737b071081/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 854.783441] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Uploaded image 2d406018-3303-46ca-b943-5ea7a0ad206c to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 854.785773] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 854.786103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-920edf6c-e83a-4cd2-a61d-73674a393a84 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.792023] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 854.792023] env[62558]: value = "task-1266810" [ 854.792023] env[62558]: _type = "Task" [ 854.792023] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.799940] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266810, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.883857] env[62558]: DEBUG nova.compute.manager [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Received event network-vif-plugged-1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.883909] env[62558]: DEBUG oslo_concurrency.lockutils [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] Acquiring lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.884205] env[62558]: DEBUG oslo_concurrency.lockutils [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] Lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.884376] env[62558]: DEBUG oslo_concurrency.lockutils [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] Lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.884559] env[62558]: DEBUG nova.compute.manager [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] No waiting events found dispatching network-vif-plugged-1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.884734] env[62558]: WARNING nova.compute.manager [req-160f6c02-e6e8-4225-869b-375b311d37c3 req-b78580b1-81ba-4f9b-b395-b5ba6d123a0e service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Received unexpected event network-vif-plugged-1288ac96-71bf-44b8-824c-f6c2a017d02d for instance with vm_state building and task_state spawning. [ 854.975798] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Successfully updated port: 1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.002979] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266809, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.052806] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.052806] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.055407] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.277s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.056872] env[62558]: INFO nova.compute.claims [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.241559] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266808, 'name': ReconfigVM_Task, 'duration_secs': 0.694341} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.241912] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfigured VM instance instance-0000003c to attach disk [datastore2] volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c/volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.246844] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a963831c-fd68-4491-a206-755bce3d0fe1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.262773] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 855.262773] env[62558]: value = "task-1266811" [ 855.262773] env[62558]: _type = "Task" [ 855.262773] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.270848] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.300789] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266810, 'name': Destroy_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.478817] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.478968] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquired lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.479149] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.501055] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266809, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.763571} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.501316] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884/a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.501586] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.501836] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38e8e914-27ae-47d6-93d8-37537d09f8a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.507972] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 855.507972] env[62558]: value = "task-1266812" [ 855.507972] env[62558]: _type = "Task" [ 855.507972] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.515494] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.562145] env[62558]: DEBUG nova.compute.utils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.565431] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.565594] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.618884] env[62558]: DEBUG nova.policy [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd44eed094d3d4e9f92bd0c99e6566986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7220b5635a34080aeaa7420b92990ad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.773032] env[62558]: DEBUG oslo_vmware.api [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266811, 'name': ReconfigVM_Task, 'duration_secs': 0.318698} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.773257] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272538', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'name': 'volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1d242e6-1561-4bd4-8e39-281ab6346661', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'serial': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 855.803377] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266810, 'name': Destroy_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.860304] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Successfully created port: b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.018428] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.023910] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077557} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.024408] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.027928] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aed3f6f-4070-4e4b-9872-9e01d423abbb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.050728] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884/a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.051691] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f017624a-14a7-4003-a326-8c75961c7ae9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.068768] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.082079] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 856.082079] env[62558]: value = "task-1266813" [ 856.082079] env[62558]: _type = "Task" [ 856.082079] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.092271] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266813, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.108696] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Successfully created port: 3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.303173] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266810, 'name': Destroy_Task, 'duration_secs': 1.170975} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.305953] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Destroyed the VM [ 856.306098] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 856.306726] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4d4a21cd-1861-426f-bea5-9d9061eb11d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.314338] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 856.314338] env[62558]: value = "task-1266814" [ 856.314338] env[62558]: _type = "Task" [ 856.314338] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.327156] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266814, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.329713] env[62558]: DEBUG nova.network.neutron [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Updating instance_info_cache with network_info: [{"id": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "address": "fa:16:3e:8c:1f:fd", "network": {"id": "3479b302-948f-44fc-a9cd-83565981ac82", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1229756164-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e75f3bcc65ee48d28da7ee016b2fd4f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1288ac96-71", "ovs_interfaceid": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.439076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f724d9-e487-481d-8a98-4ea7689309a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.446791] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f72434d-3207-4bbf-8d43-73c15b9193ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.491111] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67b3b0b-40f2-435d-8203-54dd94b3161d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.501383] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220d48a4-a5da-49a5-9229-641ac9d3b795 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.516695] env[62558]: DEBUG nova.compute.provider_tree [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.592254] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266813, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.823972] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266814, 'name': RemoveSnapshot_Task, 'duration_secs': 0.363376} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.824308] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 856.824590] env[62558]: DEBUG nova.compute.manager [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.825368] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f7686d-6708-4425-a19b-b1779a6ba51e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.833347] env[62558]: DEBUG nova.objects.instance [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.834559] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Releasing lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.834815] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Instance network_info: |[{"id": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "address": "fa:16:3e:8c:1f:fd", "network": {"id": "3479b302-948f-44fc-a9cd-83565981ac82", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1229756164-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e75f3bcc65ee48d28da7ee016b2fd4f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1288ac96-71", "ovs_interfaceid": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.835174] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:1f:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1288ac96-71bf-44b8-824c-f6c2a017d02d', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.843332] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Creating folder: Project (e75f3bcc65ee48d28da7ee016b2fd4f5). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.843931] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e631ff5-1310-49ca-bfa6-4591c63bbb06 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.854039] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Created folder: Project (e75f3bcc65ee48d28da7ee016b2fd4f5) in parent group-v272451. [ 856.854164] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Creating folder: Instances. Parent ref: group-v272540. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.855432] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06a295da-b62a-4274-84d1-574cde1ecd6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.863184] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Created folder: Instances in parent group-v272540. [ 856.863421] env[62558]: DEBUG oslo.service.loopingcall [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.863616] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.863879] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1e44825-d82a-4e24-8365-7df419ec7eb3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.882878] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.882878] env[62558]: value = "task-1266817" [ 856.882878] env[62558]: _type = "Task" [ 856.882878] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.890051] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266817, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.907152] env[62558]: DEBUG nova.compute.manager [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Received event network-changed-1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.907350] env[62558]: DEBUG nova.compute.manager [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Refreshing instance network info cache due to event network-changed-1288ac96-71bf-44b8-824c-f6c2a017d02d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.907564] env[62558]: DEBUG oslo_concurrency.lockutils [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] Acquiring lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.907710] env[62558]: DEBUG oslo_concurrency.lockutils [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] Acquired lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.907867] env[62558]: DEBUG nova.network.neutron [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Refreshing network info cache for port 1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.959279] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.020015] env[62558]: DEBUG nova.scheduler.client.report [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.080947] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.092210] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266813, 'name': ReconfigVM_Task, 'duration_secs': 0.884226} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.092918] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Reconfigured VM instance instance-00000049 to attach disk [datastore2] a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884/a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.093186] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d9b93a42-27dc-4518-a04f-781843986d18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.101357] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 857.101357] env[62558]: value = "task-1266818" [ 857.101357] env[62558]: _type = "Task" [ 857.101357] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.107336] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.107619] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.107790] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.107981] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.108183] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.108383] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.108630] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.108818] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.108996] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.109178] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.109358] env[62558]: DEBUG nova.virt.hardware [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.110125] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb85f49a-bd76-4dda-bca9-e8c438ceeded {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.115518] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266818, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.121084] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfec30b-19e0-4794-bf15-0d50d5e17419 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.348741] env[62558]: INFO nova.compute.manager [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Shelve offloading [ 857.350148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68dd179b-2049-4c27-a4ca-86cbd37baa36 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.800s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.350958] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.392s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.351160] env[62558]: DEBUG nova.compute.manager [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.351999] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4554bfe4-ea6d-48b5-9d44-f25e1b9685e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.355762] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.355762] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fce3dea5-5175-485d-8d4e-7a807c94fbdc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.360650] env[62558]: DEBUG nova.compute.manager [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 857.361193] env[62558]: DEBUG nova.objects.instance [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.364046] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 857.364046] env[62558]: value = "task-1266819" [ 857.364046] env[62558]: _type = "Task" [ 857.364046] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.372603] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 857.372603] env[62558]: DEBUG nova.compute.manager [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.373817] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fb0b23-1908-49a4-8f3b-1cb8a4a78359 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.379502] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.379683] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.379859] env[62558]: DEBUG nova.network.neutron [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.392873] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266817, 'name': CreateVM_Task, 'duration_secs': 0.44984} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.393040] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.393702] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.393864] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.394301] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.394458] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e1267fe-669a-49fc-a630-95650015d9a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.400506] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 857.400506] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5238559b-1f11-3549-1d7b-075c4cc63a48" [ 857.400506] env[62558]: _type = "Task" [ 857.400506] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.408158] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5238559b-1f11-3549-1d7b-075c4cc63a48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.525353] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.526518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.670s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.526758] env[62558]: DEBUG nova.objects.instance [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'resources' on Instance uuid e846e43d-68ec-4de3-ba62-e538643b6e4b {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.614337] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266818, 'name': Rename_Task, 'duration_secs': 0.261407} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.615129] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Successfully updated port: b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.615740] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.616341] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c773c6a5-503f-423f-ad3c-641f3c29106e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.622330] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 857.622330] env[62558]: value = "task-1266820" [ 857.622330] env[62558]: _type = "Task" [ 857.622330] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.630184] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.728545] env[62558]: DEBUG nova.network.neutron [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Updated VIF entry in instance network info cache for port 1288ac96-71bf-44b8-824c-f6c2a017d02d. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.729020] env[62558]: DEBUG nova.network.neutron [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Updating instance_info_cache with network_info: [{"id": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "address": "fa:16:3e:8c:1f:fd", "network": {"id": "3479b302-948f-44fc-a9cd-83565981ac82", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1229756164-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e75f3bcc65ee48d28da7ee016b2fd4f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1288ac96-71", "ovs_interfaceid": "1288ac96-71bf-44b8-824c-f6c2a017d02d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.867201] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.867464] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57098fec-affb-4cd5-ae6d-974f6936309a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.875550] env[62558]: DEBUG oslo_vmware.api [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 857.875550] env[62558]: value = "task-1266821" [ 857.875550] env[62558]: _type = "Task" [ 857.875550] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.889599] env[62558]: DEBUG oslo_vmware.api [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.914197] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5238559b-1f11-3549-1d7b-075c4cc63a48, 'name': SearchDatastore_Task, 'duration_secs': 0.020913} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.914639] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.914960] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.915300] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.915583] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.915868] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.916250] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f5346b4-296a-4ef2-b48b-b28270a6c0c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.925474] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.925737] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.926476] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c845f52a-e3ca-4d0d-8f90-efc2d2cd9194 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.936998] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 857.936998] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52666408-6a37-4ed2-193f-e6dd88c8b559" [ 857.936998] env[62558]: _type = "Task" [ 857.936998] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.943237] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52666408-6a37-4ed2-193f-e6dd88c8b559, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.030109] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "e1430c84-b883-419b-a180-cd76f0f09603" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.030557] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "e1430c84-b883-419b-a180-cd76f0f09603" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.125926] env[62558]: DEBUG nova.network.neutron [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updating instance_info_cache with network_info: [{"id": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "address": "fa:16:3e:20:10:d9", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eb4381-71", "ovs_interfaceid": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.137840] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266820, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.232943] env[62558]: DEBUG oslo_concurrency.lockutils [req-db2b95dc-9c46-4eb8-858a-0de5c34f7a04 req-9ff80e26-07b7-40ec-ac34-cc1de67692cb service nova] Releasing lock "refresh_cache-80221843-4e15-4f20-aeb4-4e6081371b95" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.350568] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9846b12d-b18f-441d-97b1-c6c46dfee11f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.358965] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf07933-f23c-4b80-8ca5-b7bf1c0c7e80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.391998] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215ea9b1-5414-4f56-8007-7cdda2432c07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.401538] env[62558]: DEBUG oslo_vmware.api [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266821, 'name': PowerOffVM_Task, 'duration_secs': 0.18434} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.403383] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.403577] env[62558]: DEBUG nova.compute.manager [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.404378] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26403df3-b429-4e71-80e7-b7c3a95ae1f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.407698] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6cc380-4974-4054-8a7e-f3360915cc12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.423994] env[62558]: DEBUG nova.compute.provider_tree [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.444370] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52666408-6a37-4ed2-193f-e6dd88c8b559, 'name': SearchDatastore_Task, 'duration_secs': 0.009889} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.445519] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d60b60d-3d88-49a0-a692-bac40affc240 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.451032] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 858.451032] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522c2b2e-922e-71fd-70ae-0a50c2e02dd9" [ 858.451032] env[62558]: _type = "Task" [ 858.451032] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.458641] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522c2b2e-922e-71fd-70ae-0a50c2e02dd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.535454] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "e1430c84-b883-419b-a180-cd76f0f09603" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.536129] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.633960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.635762] env[62558]: DEBUG oslo_vmware.api [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266820, 'name': PowerOnVM_Task, 'duration_secs': 0.519393} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.636299] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.636508] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Took 8.46 seconds to spawn the instance on the hypervisor. [ 858.636691] env[62558]: DEBUG nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.637737] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425b37ce-6344-4210-9bbc-b709452df3f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.907929] env[62558]: DEBUG nova.compute.manager [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received event network-vif-unplugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.908494] env[62558]: DEBUG oslo_concurrency.lockutils [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.908494] env[62558]: DEBUG oslo_concurrency.lockutils [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.908730] env[62558]: DEBUG oslo_concurrency.lockutils [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.909035] env[62558]: DEBUG nova.compute.manager [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] No waiting events found dispatching network-vif-unplugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.909035] env[62558]: WARNING nova.compute.manager [req-41244edc-f43e-47ab-809b-5bbbf0def97f req-6bc8ee7e-a7ac-4ab0-8ce5-c8aba2d57174 service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received unexpected event network-vif-unplugged-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c for instance with vm_state shelved and task_state shelving_offloading. [ 858.928376] env[62558]: DEBUG nova.scheduler.client.report [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.932322] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9acc7f67-ccc7-4d52-81f2-d90dc8111de6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.581s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.962236] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522c2b2e-922e-71fd-70ae-0a50c2e02dd9, 'name': SearchDatastore_Task, 'duration_secs': 0.008794} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.962384] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.962625] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 80221843-4e15-4f20-aeb4-4e6081371b95/80221843-4e15-4f20-aeb4-4e6081371b95.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.962898] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c3a6a42-5b79-4548-844b-f0b575aa77b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.969391] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 858.969391] env[62558]: value = "task-1266822" [ 858.969391] env[62558]: _type = "Task" [ 858.969391] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.977381] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.979522] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.980284] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93209d1e-307f-405b-86cc-6831093cf2f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.986805] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.987063] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9623173e-f912-46a7-9403-65a18ec0fdc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.042023] env[62558]: DEBUG nova.compute.utils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.044838] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 859.045639] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.047149] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.047254] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.047622] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleting the datastore file [datastore1] 96361b52-b939-4409-aaf7-ba5b3bf054e7 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.047955] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eece2118-6a2f-4b59-9ace-c2dfa73ad2d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.055087] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 859.055087] env[62558]: value = "task-1266824" [ 859.055087] env[62558]: _type = "Task" [ 859.055087] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.062985] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.104453] env[62558]: DEBUG nova.policy [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eac80dab6bac4a9f86366df07f79f380', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '899e4add94d74f8d9cf00dd451e1198a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.108684] env[62558]: DEBUG nova.compute.manager [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-vif-plugged-b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.108909] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Acquiring lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.109131] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.109355] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.110627] env[62558]: DEBUG nova.compute.manager [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] No waiting events found dispatching network-vif-plugged-b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.110627] env[62558]: WARNING nova.compute.manager [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received unexpected event network-vif-plugged-b780e62c-f04a-408f-a8e2-2e2935b57a41 for instance with vm_state building and task_state spawning. [ 859.110627] env[62558]: DEBUG nova.compute.manager [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-changed-b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.110627] env[62558]: DEBUG nova.compute.manager [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Refreshing instance network info cache due to event network-changed-b780e62c-f04a-408f-a8e2-2e2935b57a41. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.110627] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Acquiring lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.110948] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Acquired lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.110948] env[62558]: DEBUG nova.network.neutron [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Refreshing network info cache for port b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.157342] env[62558]: INFO nova.compute.manager [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Took 38.21 seconds to build instance. [ 859.435936] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.439855] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.746s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.441368] env[62558]: INFO nova.compute.claims [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.463545] env[62558]: INFO nova.scheduler.client.report [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance e846e43d-68ec-4de3-ba62-e538643b6e4b [ 859.479514] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50198} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.479786] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 80221843-4e15-4f20-aeb4-4e6081371b95/80221843-4e15-4f20-aeb4-4e6081371b95.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.479989] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.480267] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b9aa996-dc6b-46ba-bdbc-a9f11264bca9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.486753] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 859.486753] env[62558]: value = "task-1266825" [ 859.486753] env[62558]: _type = "Task" [ 859.486753] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.498439] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.549760] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.566240] env[62558]: DEBUG oslo_vmware.api [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284647} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.566552] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.566735] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.566908] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.590038] env[62558]: DEBUG nova.objects.instance [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.597721] env[62558]: INFO nova.scheduler.client.report [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance 96361b52-b939-4409-aaf7-ba5b3bf054e7 [ 859.663940] env[62558]: DEBUG nova.network.neutron [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.663940] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3d16731f-5e7e-4cfa-83c0-10048c97de89 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.328s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.677627] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Successfully created port: 8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.751662] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Successfully updated port: 3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.778593] env[62558]: DEBUG nova.network.neutron [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.971906] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7194b803-a06a-4174-9544-acab3da271aa tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "e846e43d-68ec-4de3-ba62-e538643b6e4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.525s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.999015] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067338} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.999385] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.000149] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76af69cb-6c4d-4b68-93a2-d9835d89d354 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.028467] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 80221843-4e15-4f20-aeb4-4e6081371b95/80221843-4e15-4f20-aeb4-4e6081371b95.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.029355] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa3a8436-93b8-4557-baba-291d1c6dbce5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.049696] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 860.049696] env[62558]: value = "task-1266826" [ 860.049696] env[62558]: _type = "Task" [ 860.049696] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.060648] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.096473] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.096625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.096767] env[62558]: DEBUG nova.network.neutron [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.096949] env[62558]: DEBUG nova.objects.instance [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'info_cache' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.102034] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.257201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.280987] env[62558]: DEBUG oslo_concurrency.lockutils [req-e176cc42-9c6a-4039-a2e1-026adf9f7ad8 req-db393f0a-92f3-47ad-ad13-c09ea2f5fcee service nova] Releasing lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.281408] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.281626] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.560131] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266826, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.564169] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 860.593422] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.593573] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.593621] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.594335] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.594335] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.594335] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.594518] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.594553] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.594872] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.594947] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.595132] env[62558]: DEBUG nova.virt.hardware [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.596130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae527e4e-1c7c-4f0b-b926-ff8e2d8c391f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.600429] env[62558]: DEBUG nova.objects.base [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 860.609595] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7929f146-d7e3-46d3-90e4-2b3eea2daabb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.753877] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c472d69-b04e-4dff-9d0d-99ad55374b69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.762842] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff6a857-189b-4abe-9fd8-0e039f1c6b3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.799230] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8723fcd-598f-4a3d-a3c3-923fb158ca86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.807057] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f926b88-ccd4-465f-97a2-2dbbff5a3276 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.821483] env[62558]: DEBUG nova.compute.provider_tree [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.840283] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.061769] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266826, 'name': ReconfigVM_Task, 'duration_secs': 0.754581} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.062096] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 80221843-4e15-4f20-aeb4-4e6081371b95/80221843-4e15-4f20-aeb4-4e6081371b95.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.062758] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5817d0a-c280-417b-8af1-740a80eac534 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.072515] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 861.072515] env[62558]: value = "task-1266827" [ 861.072515] env[62558]: _type = "Task" [ 861.072515] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.083150] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266827, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.163633] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.163887] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.196705] env[62558]: DEBUG nova.network.neutron [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updating instance_info_cache with network_info: [{"id": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "address": "fa:16:3e:87:61:0c", "network": {"id": "3a1c506e-ee87-4bee-b823-e97f4e775d1d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1647911419", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb780e62c-f0", "ovs_interfaceid": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3cdad4ed-33d8-46f8-9093-6904227db761", "address": "fa:16:3e:80:27:a5", "network": {"id": "32080404-692d-4426-9d0d-317667328ebf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1248140584", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cdad4ed-33", "ovs_interfaceid": "3cdad4ed-33d8-46f8-9093-6904227db761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.253211] env[62558]: DEBUG nova.compute.manager [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Received event network-changed-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.253211] env[62558]: DEBUG nova.compute.manager [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Refreshing instance network info cache due to event network-changed-37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 861.253211] env[62558]: DEBUG oslo_concurrency.lockutils [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] Acquiring lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.253211] env[62558]: DEBUG oslo_concurrency.lockutils [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] Acquired lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.253211] env[62558]: DEBUG nova.network.neutron [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Refreshing network info cache for port 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.327670] env[62558]: DEBUG nova.scheduler.client.report [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.405590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.405966] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.406491] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.407214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.407618] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.411082] env[62558]: INFO nova.compute.manager [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Terminating instance [ 861.414435] env[62558]: DEBUG nova.compute.manager [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.414642] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.415503] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0918d5d1-171f-4f26-8fd0-a26331f4ecd0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.424500] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.424758] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da8bd752-c154-4e47-916b-64f5f3fb63ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.431995] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 861.431995] env[62558]: value = "task-1266828" [ 861.431995] env[62558]: _type = "Task" [ 861.431995] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.446277] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.518669] env[62558]: DEBUG nova.compute.manager [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-vif-plugged-3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.519268] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Acquiring lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.519446] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.519688] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.519891] env[62558]: DEBUG nova.compute.manager [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] No waiting events found dispatching network-vif-plugged-3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 861.520825] env[62558]: WARNING nova.compute.manager [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received unexpected event network-vif-plugged-3cdad4ed-33d8-46f8-9093-6904227db761 for instance with vm_state building and task_state spawning. [ 861.521234] env[62558]: DEBUG nova.compute.manager [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-changed-3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.521428] env[62558]: DEBUG nova.compute.manager [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Refreshing instance network info cache due to event network-changed-3cdad4ed-33d8-46f8-9093-6904227db761. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 861.521638] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Acquiring lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.546187] env[62558]: DEBUG nova.network.neutron [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.569079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.585038] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266827, 'name': Rename_Task, 'duration_secs': 0.138706} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.585038] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.585038] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ccc6d8d-a6e6-48b7-93a1-031d259ad582 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.594020] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 861.594020] env[62558]: value = "task-1266829" [ 861.594020] env[62558]: _type = "Task" [ 861.594020] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.600480] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.669061] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.703026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Releasing lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.703026] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance network_info: |[{"id": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "address": "fa:16:3e:87:61:0c", "network": {"id": "3a1c506e-ee87-4bee-b823-e97f4e775d1d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1647911419", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb780e62c-f0", "ovs_interfaceid": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3cdad4ed-33d8-46f8-9093-6904227db761", "address": "fa:16:3e:80:27:a5", "network": {"id": "32080404-692d-4426-9d0d-317667328ebf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1248140584", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cdad4ed-33", "ovs_interfaceid": "3cdad4ed-33d8-46f8-9093-6904227db761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.703026] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Acquired lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.703026] env[62558]: DEBUG nova.network.neutron [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Refreshing network info cache for port 3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.703026] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:61:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e0cfc48-d93b-4477-8082-69a2f7aa7701', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b780e62c-f04a-408f-a8e2-2e2935b57a41', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:27:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ded8bac-871f-491b-94ec-cb67c08bc828', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cdad4ed-33d8-46f8-9093-6904227db761', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.714422] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Creating folder: Project (c7220b5635a34080aeaa7420b92990ad). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.715833] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-670dc5c0-2e06-4536-865d-72142b90805e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.729024] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Created folder: Project (c7220b5635a34080aeaa7420b92990ad) in parent group-v272451. [ 861.729024] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Creating folder: Instances. Parent ref: group-v272543. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.729024] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21ea1d3d-0c5b-4cca-8982-9b5c7324c0fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.736437] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Created folder: Instances in parent group-v272543. [ 861.736852] env[62558]: DEBUG oslo.service.loopingcall [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.737181] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.737511] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-403e0b1b-93f1-4a53-bfa2-3bb39f370eb7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.767130] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.767130] env[62558]: value = "task-1266832" [ 861.767130] env[62558]: _type = "Task" [ 861.767130] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.774925] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266832, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.778461] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Successfully updated port: 8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.835693] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.836460] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 861.839413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.271s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.841357] env[62558]: INFO nova.compute.claims [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.847386] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "016c31f6-648c-41f5-909f-5a1ae6366b98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.847625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.945948] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266828, 'name': PowerOffVM_Task, 'duration_secs': 0.257217} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.946365] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 861.946645] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.947179] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35cd9e21-295f-4d79-acb7-6c03f2df5118 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.995588] env[62558]: DEBUG nova.network.neutron [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updated VIF entry in instance network info cache for port 37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.995588] env[62558]: DEBUG nova.network.neutron [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updating instance_info_cache with network_info: [{"id": "37eb4381-7196-4cdf-bf0c-76c5d9e6fa8c", "address": "fa:16:3e:20:10:d9", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": null, "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap37eb4381-71", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.019585] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.019585] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.019585] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleting the datastore file [datastore2] 6243631f-b240-4d7c-8910-0bf3d2dedb77 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.019585] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d9f43ab-c200-40f9-b17b-7597c5fa67c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.024282] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 862.024282] env[62558]: value = "task-1266834" [ 862.024282] env[62558]: _type = "Task" [ 862.024282] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.032125] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.052555] env[62558]: DEBUG oslo_concurrency.lockutils [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.102760] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266829, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.189688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.274449] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266832, 'name': CreateVM_Task, 'duration_secs': 0.462232} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.275104] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.275529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.275529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.276479] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.276479] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ddbdfe9-fa6b-4e8d-9870-94fb0a01de11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.280628] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.280879] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquired lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.280879] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.282011] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 862.282011] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5277e0c3-ca92-03c1-df52-e4f11e0a6199" [ 862.282011] env[62558]: _type = "Task" [ 862.282011] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.290237] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5277e0c3-ca92-03c1-df52-e4f11e0a6199, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.345995] env[62558]: DEBUG nova.compute.utils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.347576] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.347697] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.352050] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.391815] env[62558]: DEBUG nova.policy [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c5503a5bf734483af68683a8c853a71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9d4de9f22ec414d90eb8c2ed9c42d18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.448657] env[62558]: DEBUG nova.network.neutron [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updated VIF entry in instance network info cache for port 3cdad4ed-33d8-46f8-9093-6904227db761. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.449124] env[62558]: DEBUG nova.network.neutron [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updating instance_info_cache with network_info: [{"id": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "address": "fa:16:3e:87:61:0c", "network": {"id": "3a1c506e-ee87-4bee-b823-e97f4e775d1d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1647911419", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e0cfc48-d93b-4477-8082-69a2f7aa7701", "external-id": "nsx-vlan-transportzone-275", "segmentation_id": 275, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb780e62c-f0", "ovs_interfaceid": "b780e62c-f04a-408f-a8e2-2e2935b57a41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3cdad4ed-33d8-46f8-9093-6904227db761", "address": "fa:16:3e:80:27:a5", "network": {"id": "32080404-692d-4426-9d0d-317667328ebf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1248140584", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cdad4ed-33", "ovs_interfaceid": "3cdad4ed-33d8-46f8-9093-6904227db761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.497609] env[62558]: DEBUG oslo_concurrency.lockutils [req-5349fa50-cc59-4bac-8eba-eb3045175cbc req-e75242be-b7d2-4c01-a82e-58ec7d507bfc service nova] Releasing lock "refresh_cache-96361b52-b939-4409-aaf7-ba5b3bf054e7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.534508] env[62558]: DEBUG oslo_vmware.api [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148924} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.534845] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 862.534979] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 862.535251] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 862.535402] env[62558]: INFO nova.compute.manager [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Took 1.12 seconds to destroy the instance on the hypervisor. [ 862.535661] env[62558]: DEBUG oslo.service.loopingcall [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.535851] env[62558]: DEBUG nova.compute.manager [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 862.535946] env[62558]: DEBUG nova.network.neutron [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 862.559091] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.559407] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96c10810-90eb-4888-8eaf-bfd60aa00618 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.568591] env[62558]: DEBUG oslo_vmware.api [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 862.568591] env[62558]: value = "task-1266835" [ 862.568591] env[62558]: _type = "Task" [ 862.568591] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.575906] env[62558]: DEBUG oslo_vmware.api [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266835, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.602278] env[62558]: DEBUG oslo_vmware.api [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266829, 'name': PowerOnVM_Task, 'duration_secs': 0.653194} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.602606] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.602805] env[62558]: INFO nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Took 7.97 seconds to spawn the instance on the hypervisor. [ 862.603199] env[62558]: DEBUG nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.604124] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27f497f-98e7-4ff8-87aa-93c1f11f8ebe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.795893] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5277e0c3-ca92-03c1-df52-e4f11e0a6199, 'name': SearchDatastore_Task, 'duration_secs': 0.009715} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.796223] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.796457] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.796729] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.796872] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.797020] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.797284] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32767167-be6d-493f-a8d9-06e67c4fa7ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.805954] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.806146] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.806844] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b890b6f-cb51-4de8-bf95-0b3e9a949a59 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.811980] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 862.811980] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f5e8d-313b-37f2-c39d-a794f7ccaeeb" [ 862.811980] env[62558]: _type = "Task" [ 862.811980] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.820052] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.826917] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525f5e8d-313b-37f2-c39d-a794f7ccaeeb, 'name': SearchDatastore_Task, 'duration_secs': 0.008768} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.827669] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce3aa722-2bbe-4680-8719-c76d3ca0632c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.832824] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 862.832824] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad7812-1598-1292-07e7-e63bbf90a49d" [ 862.832824] env[62558]: _type = "Task" [ 862.832824] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.840476] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad7812-1598-1292-07e7-e63bbf90a49d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.847245] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Successfully created port: 174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.852525] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 862.879559] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.952083] env[62558]: DEBUG oslo_concurrency.lockutils [req-c94ea471-320f-423b-a68c-d965446ad5d1 req-925496d1-7ff2-40e3-9aac-939c117a39a3 service nova] Releasing lock "refresh_cache-2b8430ef-c8eb-4eb8-a754-3c552662b966" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.982366] env[62558]: DEBUG nova.network.neutron [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Updating instance_info_cache with network_info: [{"id": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "address": "fa:16:3e:b0:27:8f", "network": {"id": "5222acbc-1bb0-4d65-b016-3dd37bd5e352", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-28747225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "899e4add94d74f8d9cf00dd451e1198a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e94d8d1-be", "ovs_interfaceid": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.080742] env[62558]: DEBUG oslo_vmware.api [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266835, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.127359] env[62558]: INFO nova.compute.manager [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Took 38.71 seconds to build instance. [ 863.218912] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fb9473-3b86-43c9-b34f-ca1ddc287bca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.227534] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a31924-5320-435b-a25f-b201917810f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.266894] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec61b13f-2b89-4461-92ba-277ad1a023e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.274522] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ac982b-5fb8-41a5-a7a5-cc76dc42f3b1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.280364] env[62558]: DEBUG nova.network.neutron [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.293958] env[62558]: DEBUG nova.compute.provider_tree [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.297196] env[62558]: DEBUG nova.compute.manager [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Received event network-vif-plugged-8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.297462] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Acquiring lock "574525eb-0535-4664-8449-813c16e4781c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.297704] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Lock "574525eb-0535-4664-8449-813c16e4781c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.297874] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Lock "574525eb-0535-4664-8449-813c16e4781c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.298281] env[62558]: DEBUG nova.compute.manager [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] No waiting events found dispatching network-vif-plugged-8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.298479] env[62558]: WARNING nova.compute.manager [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Received unexpected event network-vif-plugged-8e94d8d1-be4c-4543-aae7-d521cff11fea for instance with vm_state building and task_state spawning. [ 863.298666] env[62558]: DEBUG nova.compute.manager [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Received event network-changed-8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.298838] env[62558]: DEBUG nova.compute.manager [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Refreshing instance network info cache due to event network-changed-8e94d8d1-be4c-4543-aae7-d521cff11fea. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.299018] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Acquiring lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.343079] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad7812-1598-1292-07e7-e63bbf90a49d, 'name': SearchDatastore_Task, 'duration_secs': 0.009728} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.343351] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.343625] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2b8430ef-c8eb-4eb8-a754-3c552662b966/2b8430ef-c8eb-4eb8-a754-3c552662b966.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.343883] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-248ed212-8a70-46fa-a9db-7385caca890c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.349896] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 863.349896] env[62558]: value = "task-1266836" [ 863.349896] env[62558]: _type = "Task" [ 863.349896] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.357628] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266836, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.484963] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Releasing lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.484963] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Instance network_info: |[{"id": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "address": "fa:16:3e:b0:27:8f", "network": {"id": "5222acbc-1bb0-4d65-b016-3dd37bd5e352", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-28747225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "899e4add94d74f8d9cf00dd451e1198a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e94d8d1-be", "ovs_interfaceid": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.485775] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Acquired lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.485775] env[62558]: DEBUG nova.network.neutron [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Refreshing network info cache for port 8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.487192] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:27:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e94d8d1-be4c-4543-aae7-d521cff11fea', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.495176] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Creating folder: Project (899e4add94d74f8d9cf00dd451e1198a). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.498112] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23e04d92-5d45-41f4-9892-c09c79b1a27e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.508818] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Created folder: Project (899e4add94d74f8d9cf00dd451e1198a) in parent group-v272451. [ 863.509045] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Creating folder: Instances. Parent ref: group-v272546. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.509305] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d773ee08-0723-49b2-8c4c-4155e12c03de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.518890] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Created folder: Instances in parent group-v272546. [ 863.519138] env[62558]: DEBUG oslo.service.loopingcall [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.519351] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.519576] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17c30d7f-1dbf-4ce5-90fc-760b502c712b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.545067] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.545067] env[62558]: value = "task-1266839" [ 863.545067] env[62558]: _type = "Task" [ 863.545067] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.553556] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266839, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.586175] env[62558]: DEBUG oslo_vmware.api [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266835, 'name': PowerOnVM_Task, 'duration_secs': 0.711927} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.587958] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.588664] env[62558]: DEBUG nova.compute.manager [None req-46b55cdb-b519-4ce8-a5fd-3e79ad933e86 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.591017] env[62558]: DEBUG nova.compute.manager [req-73a718ef-7bd4-4473-a4b9-a05a923128d7 req-ceec9d65-eaf7-4d46-924c-9dd93af9a3c6 service nova] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Received event network-vif-deleted-fbdcb890-4930-4b7e-8c5c-12082038f0cf {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.591949] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c4bfc7-f80f-45c4-90e2-d798ff752861 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.629162] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3b661a0e-e38a-4eb1-b1c6-e3d6010e9ceb tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.026s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.758279] env[62558]: DEBUG nova.network.neutron [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Updated VIF entry in instance network info cache for port 8e94d8d1-be4c-4543-aae7-d521cff11fea. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.758567] env[62558]: DEBUG nova.network.neutron [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Updating instance_info_cache with network_info: [{"id": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "address": "fa:16:3e:b0:27:8f", "network": {"id": "5222acbc-1bb0-4d65-b016-3dd37bd5e352", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-28747225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "899e4add94d74f8d9cf00dd451e1198a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e94d8d1-be", "ovs_interfaceid": "8e94d8d1-be4c-4543-aae7-d521cff11fea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.784031] env[62558]: INFO nova.compute.manager [-] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Took 1.25 seconds to deallocate network for instance. [ 863.804172] env[62558]: DEBUG nova.scheduler.client.report [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.859844] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266836, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459242} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.860253] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 2b8430ef-c8eb-4eb8-a754-3c552662b966/2b8430ef-c8eb-4eb8-a754-3c552662b966.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.860351] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.860586] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c93a9288-1c4a-47c0-99de-0b4576f92149 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.864708] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 863.867721] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 863.867721] env[62558]: value = "task-1266840" [ 863.867721] env[62558]: _type = "Task" [ 863.867721] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.878049] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.892891] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.893193] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.893356] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.893539] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.893688] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.893840] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.894064] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.894231] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.894400] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.894750] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.894750] env[62558]: DEBUG nova.virt.hardware [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.895605] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f902a2-e86e-4d20-8fb4-1400d84f1a9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.905651] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3082f1-880c-4239-8fe1-7b83feee47d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.055087] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266839, 'name': CreateVM_Task, 'duration_secs': 0.50087} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.055268] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.055966] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.056166] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.056488] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.056774] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-785b1526-0016-4056-b445-932616ceaa98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.061753] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 864.061753] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b1c246-1b4a-e35f-a428-e8cbe81ce49e" [ 864.061753] env[62558]: _type = "Task" [ 864.061753] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.070128] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b1c246-1b4a-e35f-a428-e8cbe81ce49e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.262911] env[62558]: DEBUG oslo_concurrency.lockutils [req-bfe9fcac-4827-4c3a-93d6-6e98b96f807d req-3e4d95cb-ee1a-499a-ae0b-bc18376e218a service nova] Releasing lock "refresh_cache-574525eb-0535-4664-8449-813c16e4781c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.291308] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.312381] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.312958] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.316383] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.377s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.316682] env[62558]: DEBUG nova.objects.instance [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lazy-loading 'resources' on Instance uuid 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.319590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.319590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.319590] env[62558]: INFO nova.compute.manager [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Shelving [ 864.378631] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074718} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.378989] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.379815] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9aafd9-b2f1-40e2-8fd5-014a11ea0803 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.408267] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 2b8430ef-c8eb-4eb8-a754-3c552662b966/2b8430ef-c8eb-4eb8-a754-3c552662b966.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.408679] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd2bc27a-4343-4dc2-9791-dc0e514a1314 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.430894] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 864.430894] env[62558]: value = "task-1266841" [ 864.430894] env[62558]: _type = "Task" [ 864.430894] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.442217] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266841, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.572455] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b1c246-1b4a-e35f-a428-e8cbe81ce49e, 'name': SearchDatastore_Task, 'duration_secs': 0.009088} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.572802] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.573056] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.573298] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.573448] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.573623] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.573884] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3823f54d-20ad-4235-b980-9bcaa9354609 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.581760] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.581954] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.582675] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3b579a1-4493-4759-b335-6ea9c800193d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.587881] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 864.587881] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ee8381-9156-5e06-f61f-934e0047acd5" [ 864.587881] env[62558]: _type = "Task" [ 864.587881] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.596894] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ee8381-9156-5e06-f61f-934e0047acd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.653262] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Successfully updated port: 174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.823379] env[62558]: DEBUG nova.compute.utils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.830789] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.831012] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 864.836168] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.836816] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0842e77c-4e25-41a5-bd89-68fe5305f5d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.845664] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 864.845664] env[62558]: value = "task-1266842" [ 864.845664] env[62558]: _type = "Task" [ 864.845664] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.858341] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.925096] env[62558]: DEBUG nova.policy [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd3a894e5784481eaff0fdd0fa414cee', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b9d93976e6045c288ad6b152304ef72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.942433] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.102345] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ee8381-9156-5e06-f61f-934e0047acd5, 'name': SearchDatastore_Task, 'duration_secs': 0.009847} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.108141] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20dcc099-5e63-44a6-9996-b837a4fdfa3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.113552] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 865.113552] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52491d53-46e8-c074-6d48-0cafb58deb2b" [ 865.113552] env[62558]: _type = "Task" [ 865.113552] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.122444] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52491d53-46e8-c074-6d48-0cafb58deb2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.155966] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.156250] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.156337] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.204692] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bd8cbc-ac12-4fdd-b44c-20e3df2ad040 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.212813] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597fd75a-e0b4-4894-9751-7bf749b74dad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.246542] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cafeef-30c7-4123-806c-30af7e526570 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.254520] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ad6123-7281-4dd1-a577-eabfa91afcc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.268384] env[62558]: DEBUG nova.compute.provider_tree [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.331428] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.356968] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266842, 'name': PowerOffVM_Task, 'duration_secs': 0.214378} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.357580] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.358419] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f8d47f-b8b3-4638-b576-b436dd1ef4bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.387666] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58576c1-ab21-4089-b029-fa0ec5e5ac24 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.445020] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266841, 'name': ReconfigVM_Task, 'duration_secs': 0.596425} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.445539] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 2b8430ef-c8eb-4eb8-a754-3c552662b966/2b8430ef-c8eb-4eb8-a754-3c552662b966.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.446291] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72927f6a-c062-4938-9d60-d5eda69cb284 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.453382] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 865.453382] env[62558]: value = "task-1266843" [ 865.453382] env[62558]: _type = "Task" [ 865.453382] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.463038] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266843, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.581157] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Successfully created port: 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.605920] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "80221843-4e15-4f20-aeb4-4e6081371b95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.606290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.606557] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.606824] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.607094] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.609437] env[62558]: INFO nova.compute.manager [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Terminating instance [ 865.611298] env[62558]: DEBUG nova.compute.manager [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.611503] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.612364] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ec163d-06af-423a-b3de-2305be67cce8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.617515] env[62558]: DEBUG nova.compute.manager [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Received event network-vif-plugged-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.617711] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.617921] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.618156] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.618289] env[62558]: DEBUG nova.compute.manager [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] No waiting events found dispatching network-vif-plugged-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.618454] env[62558]: WARNING nova.compute.manager [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Received unexpected event network-vif-plugged-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 for instance with vm_state building and task_state spawning. [ 865.618625] env[62558]: DEBUG nova.compute.manager [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Received event network-changed-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.618791] env[62558]: DEBUG nova.compute.manager [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Refreshing instance network info cache due to event network-changed-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.618965] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Acquiring lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.626581] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.631800] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c334cb14-f9cc-49ee-b497-69cd2c412108 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.633714] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52491d53-46e8-c074-6d48-0cafb58deb2b, 'name': SearchDatastore_Task, 'duration_secs': 0.043192} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.633967] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.634234] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 574525eb-0535-4664-8449-813c16e4781c/574525eb-0535-4664-8449-813c16e4781c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.634835] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6dc4e8d-1a04-4981-a510-86202fbbfeb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.638682] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 865.638682] env[62558]: value = "task-1266844" [ 865.638682] env[62558]: _type = "Task" [ 865.638682] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.643926] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 865.643926] env[62558]: value = "task-1266845" [ 865.643926] env[62558]: _type = "Task" [ 865.643926] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.651715] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266844, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.656609] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.694194] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.771994] env[62558]: DEBUG nova.scheduler.client.report [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.865621] env[62558]: DEBUG nova.network.neutron [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.903873] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 865.904379] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8ef18efe-3528-4d61-89e5-9b2a7ef37a02 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.914880] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 865.914880] env[62558]: value = "task-1266846" [ 865.914880] env[62558]: _type = "Task" [ 865.914880] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.924980] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266846, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.968725] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266843, 'name': Rename_Task, 'duration_secs': 0.169433} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.969220] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.969846] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e1eb3ed-2a29-4349-a681-c7c6cbe949bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.978899] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 865.978899] env[62558]: value = "task-1266847" [ 865.978899] env[62558]: _type = "Task" [ 865.978899] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.996784] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266847, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.152036] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266844, 'name': PowerOffVM_Task, 'duration_secs': 0.331255} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.152760] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.153016] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.153285] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c464c0e-0826-4a94-aa38-1b070a969626 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.158093] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266845, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470029} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.158668] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 574525eb-0535-4664-8449-813c16e4781c/574525eb-0535-4664-8449-813c16e4781c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.158927] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.159228] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef9f7896-f423-4e87-87e5-70a07eafb574 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.164606] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 866.164606] env[62558]: value = "task-1266849" [ 866.164606] env[62558]: _type = "Task" [ 866.164606] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.172995] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266849, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.219298] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.219533] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.219766] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Deleting the datastore file [datastore2] 80221843-4e15-4f20-aeb4-4e6081371b95 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.220101] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed1771fe-2173-4faf-84b7-0a2d91fd9426 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.226797] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for the task: (returnval){ [ 866.226797] env[62558]: value = "task-1266850" [ 866.226797] env[62558]: _type = "Task" [ 866.226797] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.234783] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.280069] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.963s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.282652] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.552s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.284335] env[62558]: INFO nova.compute.claims [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.304134] env[62558]: INFO nova.scheduler.client.report [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Deleted allocations for instance 867c0fb3-d4f6-4945-a5cc-6d93efcadb79 [ 866.343765] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.370031] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.370352] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.370518] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.370704] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.370854] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.371009] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.371221] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.371385] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.371676] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.371755] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.371938] env[62558]: DEBUG nova.virt.hardware [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.372665] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.372976] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Instance network_info: |[{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.373816] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c5b8d6-c6f6-43f7-8263-c34b9ea127f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.376484] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Acquired lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.376691] env[62558]: DEBUG nova.network.neutron [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Refreshing network info cache for port 174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.380911] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:76:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f256cfee-512d-4192-9aca-6750fdb1cd4c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '174e00d0-f0e5-4b0a-9ca2-4db08bb7f674', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.388277] env[62558]: DEBUG oslo.service.loopingcall [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.389405] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.390145] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a46487b-1429-4b19-86cb-506aad6dc392 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.410166] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a496d1d-1ea6-4fe9-951d-9a348a3864c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.415412] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.415412] env[62558]: value = "task-1266851" [ 866.415412] env[62558]: _type = "Task" [ 866.415412] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.438805] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266846, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.441068] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266851, 'name': CreateVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.490754] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266847, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.675317] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266849, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104274} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.675644] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.676460] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d73ba6-fd6c-42cf-bd22-63b90ac4a2fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.700561] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 574525eb-0535-4664-8449-813c16e4781c/574525eb-0535-4664-8449-813c16e4781c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.700951] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ff9b2dd-fae0-42cc-a103-4a3e49a87a09 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.722591] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 866.722591] env[62558]: value = "task-1266852" [ 866.722591] env[62558]: _type = "Task" [ 866.722591] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.733654] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.738644] env[62558]: DEBUG oslo_vmware.api [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Task: {'id': task-1266850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414674} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.738927] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.739149] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.739334] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.739525] env[62558]: INFO nova.compute.manager [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Took 1.13 seconds to destroy the instance on the hypervisor. [ 866.739794] env[62558]: DEBUG oslo.service.loopingcall [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.740055] env[62558]: DEBUG nova.compute.manager [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.740155] env[62558]: DEBUG nova.network.neutron [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.812249] env[62558]: DEBUG oslo_concurrency.lockutils [None req-83fcedd1-3528-4e3a-835c-d2f08e163e86 tempest-ListServerFiltersTestJSON-1694933546 tempest-ListServerFiltersTestJSON-1694933546-project-member] Lock "867c0fb3-d4f6-4945-a5cc-6d93efcadb79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.880s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.937008] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266846, 'name': CreateSnapshot_Task, 'duration_secs': 0.924156} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.947012] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 866.947494] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266851, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.953159] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15734271-6625-456f-a429-d9aac77961a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.993521] env[62558]: DEBUG oslo_vmware.api [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266847, 'name': PowerOnVM_Task, 'duration_secs': 0.755799} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.993973] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.994346] env[62558]: INFO nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Took 9.91 seconds to spawn the instance on the hypervisor. [ 866.994785] env[62558]: DEBUG nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.995984] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac0a4b5-039e-4111-bede-e5fa603bb31f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.087725] env[62558]: DEBUG nova.compute.manager [req-20b73467-107a-4bd0-9af3-cfadff808697 req-2f947ab6-a917-4fd6-9708-8eb28b3e8914 service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Received event network-vif-deleted-1288ac96-71bf-44b8-824c-f6c2a017d02d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.087960] env[62558]: INFO nova.compute.manager [req-20b73467-107a-4bd0-9af3-cfadff808697 req-2f947ab6-a917-4fd6-9708-8eb28b3e8914 service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Neutron deleted interface 1288ac96-71bf-44b8-824c-f6c2a017d02d; detaching it from the instance and deleting it from the info cache [ 867.090852] env[62558]: DEBUG nova.network.neutron [req-20b73467-107a-4bd0-9af3-cfadff808697 req-2f947ab6-a917-4fd6-9708-8eb28b3e8914 service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.171229] env[62558]: DEBUG nova.network.neutron [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updated VIF entry in instance network info cache for port 174e00d0-f0e5-4b0a-9ca2-4db08bb7f674. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.171739] env[62558]: DEBUG nova.network.neutron [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.233413] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266852, 'name': ReconfigVM_Task, 'duration_secs': 0.350188} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.233694] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 574525eb-0535-4664-8449-813c16e4781c/574525eb-0535-4664-8449-813c16e4781c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.234345] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08c55c51-cbf5-41a8-99b3-5a991f51df71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.240877] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 867.240877] env[62558]: value = "task-1266853" [ 867.240877] env[62558]: _type = "Task" [ 867.240877] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.249420] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266853, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.430017] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266851, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.480662] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 867.483961] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a5060b0c-0652-4808-a7c0-6b46f6620373 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.494131] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 867.494131] env[62558]: value = "task-1266854" [ 867.494131] env[62558]: _type = "Task" [ 867.494131] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.507619] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266854, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.523881] env[62558]: INFO nova.compute.manager [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Took 38.15 seconds to build instance. [ 867.557923] env[62558]: DEBUG nova.network.neutron [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.589834] env[62558]: DEBUG nova.compute.manager [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-vif-plugged-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.590903] env[62558]: DEBUG oslo_concurrency.lockutils [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.591278] env[62558]: DEBUG oslo_concurrency.lockutils [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] Lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.592482] env[62558]: DEBUG oslo_concurrency.lockutils [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] Lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.592482] env[62558]: DEBUG nova.compute.manager [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] No waiting events found dispatching network-vif-plugged-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.592482] env[62558]: WARNING nova.compute.manager [req-86040b65-f666-4e39-854b-9aba7da4f313 req-a5bc0903-1120-4c0f-bee1-1f3aa00fd7e7 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received unexpected event network-vif-plugged-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 for instance with vm_state building and task_state spawning. [ 867.596967] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Successfully updated port: 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.597407] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68f3e8e5-5e7b-4fa5-b0fb-e77533a745d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.611785] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfe5258-0519-4a7c-a15c-150591c69585 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.652820] env[62558]: DEBUG nova.compute.manager [req-20b73467-107a-4bd0-9af3-cfadff808697 req-2f947ab6-a917-4fd6-9708-8eb28b3e8914 service nova] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Detach interface failed, port_id=1288ac96-71bf-44b8-824c-f6c2a017d02d, reason: Instance 80221843-4e15-4f20-aeb4-4e6081371b95 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 867.675675] env[62558]: DEBUG oslo_concurrency.lockutils [req-2410100b-9015-4d00-a70d-31f9a76889ab req-982c2b26-c0ca-4b43-aaec-5226fe0676ec service nova] Releasing lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.736801] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c41fd5c-3bfe-49b2-bd8d-a71cdece08c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.749878] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bd112b-0722-4899-88d5-42179c6619cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.758486] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266853, 'name': Rename_Task, 'duration_secs': 0.147155} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.759244] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.759542] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfa51a8c-11ca-4586-80f6-00fdbac5163b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.800874] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec9bbc1-d9e8-44d6-b347-e76b20c30d04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.804755] env[62558]: DEBUG nova.compute.manager [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.804755] env[62558]: DEBUG nova.compute.manager [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing instance network info cache due to event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.804974] env[62558]: DEBUG oslo_concurrency.lockutils [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.805111] env[62558]: DEBUG oslo_concurrency.lockutils [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.805281] env[62558]: DEBUG nova.network.neutron [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.807908] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 867.807908] env[62558]: value = "task-1266855" [ 867.807908] env[62558]: _type = "Task" [ 867.807908] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.815607] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527ed2a9-5931-4dd5-b007-3b3adda14975 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.823147] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266855, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.833454] env[62558]: DEBUG nova.compute.provider_tree [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.926912] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266851, 'name': CreateVM_Task, 'duration_secs': 1.487592} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.927129] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.928261] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.928455] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.928973] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.929238] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97a235d3-12be-4803-a721-e04150c41c7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.934806] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 867.934806] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e1819-0ede-e3f7-a366-b57f57189ed3" [ 867.934806] env[62558]: _type = "Task" [ 867.934806] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.944089] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e1819-0ede-e3f7-a366-b57f57189ed3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.003526] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266854, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.029311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fcfc35c-4e96-49c6-8c47-c87427903aa8 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.033s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.064163] env[62558]: INFO nova.compute.manager [-] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Took 1.32 seconds to deallocate network for instance. [ 868.106084] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.320446] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266855, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.336696] env[62558]: DEBUG nova.scheduler.client.report [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.356788] env[62558]: DEBUG nova.network.neutron [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.445180] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e1819-0ede-e3f7-a366-b57f57189ed3, 'name': SearchDatastore_Task, 'duration_secs': 0.010451} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.445387] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.445570] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.445752] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.445860] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.446087] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.446359] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-132b0d25-5acb-421a-8eed-529b79a2d17f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.449532] env[62558]: DEBUG nova.network.neutron [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.468226] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.468421] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.469198] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd6a86a2-e5f9-4b2a-be43-dd965a6a555c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.474335] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 868.474335] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b66ecd-7a85-cc00-b6bc-bd4c7fd2c0d7" [ 868.474335] env[62558]: _type = "Task" [ 868.474335] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.481977] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b66ecd-7a85-cc00-b6bc-bd4c7fd2c0d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.502286] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266854, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.571317] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.632841] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.633163] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.633434] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.633707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.633960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.636263] env[62558]: INFO nova.compute.manager [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Terminating instance [ 868.637941] env[62558]: DEBUG nova.compute.manager [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.638155] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.639032] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3a5fb7-5c03-4c07-b9bd-abf014ad9587 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.646477] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.646714] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6f69f5f-6017-4b39-8c0b-63a91b4298b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.653418] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 868.653418] env[62558]: value = "task-1266856" [ 868.653418] env[62558]: _type = "Task" [ 868.653418] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.661342] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.820656] env[62558]: DEBUG oslo_vmware.api [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266855, 'name': PowerOnVM_Task, 'duration_secs': 0.826718} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.820966] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.821228] env[62558]: INFO nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Took 8.26 seconds to spawn the instance on the hypervisor. [ 868.821416] env[62558]: DEBUG nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.822204] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4193920a-87a6-4253-9526-e3853362f0a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.842305] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.560s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.842842] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.845443] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.373s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.845666] env[62558]: DEBUG nova.objects.instance [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lazy-loading 'resources' on Instance uuid 8acd06fc-b040-468c-980f-6e17b2343c4a {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.952380] env[62558]: DEBUG oslo_concurrency.lockutils [req-e48e2fe5-5af4-46dc-8a9b-fb45fe3ee6d6 req-3a49a878-46c7-48e4-8440-78536892f5b3 service nova] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.952772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.952933] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.985837] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b66ecd-7a85-cc00-b6bc-bd4c7fd2c0d7, 'name': SearchDatastore_Task, 'duration_secs': 0.032752} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.985837] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ed115f-16bf-4f52-b045-af1a689d667b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.990572] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 868.990572] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5294113c-9c67-cbb8-fe6a-23b6e8683e94" [ 868.990572] env[62558]: _type = "Task" [ 868.990572] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.000777] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5294113c-9c67-cbb8-fe6a-23b6e8683e94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.005798] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266854, 'name': CloneVM_Task, 'duration_secs': 1.392475} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.006054] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Created linked-clone VM from snapshot [ 869.006748] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2268bc-1b9a-40ce-bb7c-0d9e416cf8cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.013527] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Uploading image 2f0f0be9-1e05-4998-a402-10590e152a59 {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 869.039440] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 869.039440] env[62558]: value = "vm-272551" [ 869.039440] env[62558]: _type = "VirtualMachine" [ 869.039440] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 869.039848] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-401b4e78-ec06-43f5-ab6a-608952875fce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.047241] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lease: (returnval){ [ 869.047241] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52427c4b-0e5a-cf42-7db3-c0c7ebce4316" [ 869.047241] env[62558]: _type = "HttpNfcLease" [ 869.047241] env[62558]: } obtained for exporting VM: (result){ [ 869.047241] env[62558]: value = "vm-272551" [ 869.047241] env[62558]: _type = "VirtualMachine" [ 869.047241] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 869.047523] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the lease: (returnval){ [ 869.047523] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52427c4b-0e5a-cf42-7db3-c0c7ebce4316" [ 869.047523] env[62558]: _type = "HttpNfcLease" [ 869.047523] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 869.054294] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.054294] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52427c4b-0e5a-cf42-7db3-c0c7ebce4316" [ 869.054294] env[62558]: _type = "HttpNfcLease" [ 869.054294] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 869.167806] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266856, 'name': PowerOffVM_Task, 'duration_secs': 0.283722} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.168244] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.168568] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.168827] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8723b421-79fa-4ed0-bb39-91ce7cb68ecd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.316806] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.317216] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.317453] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Deleting the datastore file [datastore1] 2b8430ef-c8eb-4eb8-a754-3c552662b966 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.317756] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8ef2afe-0d58-49b2-8dda-f36d591cbd57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.324239] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for the task: (returnval){ [ 869.324239] env[62558]: value = "task-1266859" [ 869.324239] env[62558]: _type = "Task" [ 869.324239] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.336701] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266859, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.340576] env[62558]: INFO nova.compute.manager [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Took 38.58 seconds to build instance. [ 869.348668] env[62558]: DEBUG nova.compute.utils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.353323] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 869.353323] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.395811] env[62558]: DEBUG nova.policy [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9ea7dcf145f4d4098ee057593c1ebe7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7663e884862d43b49eabe6396d8cd107', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.497515] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.504824] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5294113c-9c67-cbb8-fe6a-23b6e8683e94, 'name': SearchDatastore_Task, 'duration_secs': 0.023149} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.505150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.505460] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.505773] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba46443f-35fe-46ea-8cf9-0897ac28d221 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.512964] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 869.512964] env[62558]: value = "task-1266860" [ 869.512964] env[62558]: _type = "Task" [ 869.512964] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.525316] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.558349] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.558349] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52427c4b-0e5a-cf42-7db3-c0c7ebce4316" [ 869.558349] env[62558]: _type = "HttpNfcLease" [ 869.558349] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 869.559041] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 869.559041] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52427c4b-0e5a-cf42-7db3-c0c7ebce4316" [ 869.559041] env[62558]: _type = "HttpNfcLease" [ 869.559041] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 869.559330] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7c7dd0-4a0f-41b1-be3b-c837ac125759 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.566707] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 869.566895] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 869.669763] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bbf900d0-d994-47f5-922e-48b4bebd3ad1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.736174] env[62558]: DEBUG nova.network.neutron [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.744302] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f84efd8-4695-49a1-9662-a5eec4bab9ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.758696] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee16bafa-eb8b-4bc9-95dd-8b44c481b7a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.796416] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Successfully created port: ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.799997] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68d6ef8-fbfc-4f78-aab4-ff538ba0a020 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.811051] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0782b85-f0e2-4f60-bc02-103ed37d166e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.826016] env[62558]: DEBUG nova.compute.provider_tree [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.837750] env[62558]: DEBUG oslo_vmware.api [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Task: {'id': task-1266859, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180581} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.838724] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.838919] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.839127] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.839306] env[62558]: INFO nova.compute.manager [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Took 1.20 seconds to destroy the instance on the hypervisor. [ 869.839556] env[62558]: DEBUG oslo.service.loopingcall [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.840075] env[62558]: DEBUG nova.compute.manager [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.840183] env[62558]: DEBUG nova.network.neutron [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.842040] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a3c8eeca-5243-4e05-941a-baca3b90503c tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.606s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.854185] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.028668] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486787} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.029160] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.029309] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.029471] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6961d46-fd6e-4a41-8044-513523b992f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.036413] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 870.036413] env[62558]: value = "task-1266861" [ 870.036413] env[62558]: _type = "Task" [ 870.036413] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.048050] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.243545] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.243545] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance network_info: |[{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.243545] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:33:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7dba3d4c-c3d7-4a6b-b03b-df7f321245c1', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.252084] env[62558]: DEBUG oslo.service.loopingcall [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.252536] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.252912] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-379badc2-26a9-4997-a79d-e612a7d9d01a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.280752] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.280752] env[62558]: value = "task-1266862" [ 870.280752] env[62558]: _type = "Task" [ 870.280752] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.291660] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266862, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.331487] env[62558]: DEBUG nova.scheduler.client.report [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.350818] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "574525eb-0535-4664-8449-813c16e4781c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.351817] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.351898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "574525eb-0535-4664-8449-813c16e4781c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.352221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.352416] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.355624] env[62558]: INFO nova.compute.manager [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Terminating instance [ 870.357462] env[62558]: DEBUG nova.compute.manager [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.357661] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.366271] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95aabd6-45c1-40fe-bbdb-4f24b39f3107 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.383197] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.383586] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fcbb32f-bb45-4dcb-a193-418be6fe04ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.391380] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 870.391380] env[62558]: value = "task-1266863" [ 870.391380] env[62558]: _type = "Task" [ 870.391380] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.406166] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.486860] env[62558]: DEBUG nova.compute.manager [req-b0ae26ec-152e-4233-813b-246fcd658025 req-93cfed37-03e7-4aa1-a522-edfb0a5b379a service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-vif-deleted-b780e62c-f04a-408f-a8e2-2e2935b57a41 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.486860] env[62558]: INFO nova.compute.manager [req-b0ae26ec-152e-4233-813b-246fcd658025 req-93cfed37-03e7-4aa1-a522-edfb0a5b379a service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Neutron deleted interface b780e62c-f04a-408f-a8e2-2e2935b57a41; detaching it from the instance and deleting it from the info cache [ 870.487409] env[62558]: DEBUG nova.network.neutron [req-b0ae26ec-152e-4233-813b-246fcd658025 req-93cfed37-03e7-4aa1-a522-edfb0a5b379a service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updating instance_info_cache with network_info: [{"id": "3cdad4ed-33d8-46f8-9093-6904227db761", "address": "fa:16:3e:80:27:a5", "network": {"id": "32080404-692d-4426-9d0d-317667328ebf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1248140584", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c7220b5635a34080aeaa7420b92990ad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ded8bac-871f-491b-94ec-cb67c08bc828", "external-id": "nsx-vlan-transportzone-212", "segmentation_id": 212, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cdad4ed-33", "ovs_interfaceid": "3cdad4ed-33d8-46f8-9093-6904227db761", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.552214] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085873} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.552214] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.552214] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139532b6-0cdd-4dec-8913-bba1f659b278 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.578295] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.579684] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76efd010-0d03-4c3c-8b6c-ff1eba969b6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.602517] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 870.602517] env[62558]: value = "task-1266864" [ 870.602517] env[62558]: _type = "Task" [ 870.602517] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.612038] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.793846] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266862, 'name': CreateVM_Task, 'duration_secs': 0.463615} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.794024] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.794733] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.794901] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.795244] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.795500] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fdc16b1-0d1f-4280-a02a-7c2b6704f2bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.800947] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 870.800947] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5285342a-a165-1ab5-d4c8-dfe9385dd816" [ 870.800947] env[62558]: _type = "Task" [ 870.800947] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.810508] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5285342a-a165-1ab5-d4c8-dfe9385dd816, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.837513] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.840389] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.649s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.840905] env[62558]: DEBUG nova.objects.instance [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lazy-loading 'resources' on Instance uuid 66ff3b25-d49c-4129-8b51-5338e75b09b1 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.859598] env[62558]: INFO nova.scheduler.client.report [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Deleted allocations for instance 8acd06fc-b040-468c-980f-6e17b2343c4a [ 870.874480] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.898109] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.898532] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.898731] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.899340] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.899504] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.899661] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.899883] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.900066] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.900244] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.900409] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.900608] env[62558]: DEBUG nova.virt.hardware [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.901463] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ef7e46-caa1-4914-8682-0ed26ff054a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.907732] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266863, 'name': PowerOffVM_Task, 'duration_secs': 0.267559} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.908510] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.909075] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.909075] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b4d2754-cf6a-4626-9683-fe7fcf5fa638 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.914714] env[62558]: DEBUG nova.network.neutron [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.920451] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa7dbbb-700c-449b-975f-7d3767142c79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.977824] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.978175] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.978412] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Deleting the datastore file [datastore1] 574525eb-0535-4664-8449-813c16e4781c {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.978761] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21f33e5f-79c1-4f8d-8236-b4baaf54cb24 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.985182] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for the task: (returnval){ [ 870.985182] env[62558]: value = "task-1266866" [ 870.985182] env[62558]: _type = "Task" [ 870.985182] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.994538] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.994911] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-129c604b-b956-4765-bbe6-f899ae5d5385 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.004492] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60126d5b-8c04-4ba9-ace5-eef6fdd5b631 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.034306] env[62558]: DEBUG nova.compute.manager [req-b0ae26ec-152e-4233-813b-246fcd658025 req-93cfed37-03e7-4aa1-a522-edfb0a5b379a service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Detach interface failed, port_id=b780e62c-f04a-408f-a8e2-2e2935b57a41, reason: Instance 2b8430ef-c8eb-4eb8-a754-3c552662b966 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 871.114426] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266864, 'name': ReconfigVM_Task, 'duration_secs': 0.398747} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.114855] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.115664] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27049929-0567-4bc5-b91e-ee71e29d5f22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.122546] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 871.122546] env[62558]: value = "task-1266867" [ 871.122546] env[62558]: _type = "Task" [ 871.122546] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.132222] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266867, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.268929] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.269448] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.269966] env[62558]: DEBUG nova.objects.instance [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid 4526df2a-f801-4de8-8218-497a7d22034f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.315198] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5285342a-a165-1ab5-d4c8-dfe9385dd816, 'name': SearchDatastore_Task, 'duration_secs': 0.017415} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.315850] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.316275] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.316744] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.317091] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.317473] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.317936] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36c3868c-4d8e-45ed-9963-2ae38f56f0d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.328717] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.328912] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.329693] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56801a8d-a5a6-4896-9166-2c38561bcc54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.335306] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 871.335306] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5213f423-579b-edb0-be83-54baef25aa16" [ 871.335306] env[62558]: _type = "Task" [ 871.335306] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.345202] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5213f423-579b-edb0-be83-54baef25aa16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.368483] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3189905a-564a-4665-a177-e596cb279102 tempest-ImagesOneServerTestJSON-1260803904 tempest-ImagesOneServerTestJSON-1260803904-project-member] Lock "8acd06fc-b040-468c-980f-6e17b2343c4a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.971s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.417588] env[62558]: INFO nova.compute.manager [-] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Took 1.58 seconds to deallocate network for instance. [ 871.503026] env[62558]: DEBUG oslo_vmware.api [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Task: {'id': task-1266866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.395482} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.504008] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.504238] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.504428] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.505037] env[62558]: INFO nova.compute.manager [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] [instance: 574525eb-0535-4664-8449-813c16e4781c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 871.505037] env[62558]: DEBUG oslo.service.loopingcall [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.505169] env[62558]: DEBUG nova.compute.manager [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.505283] env[62558]: DEBUG nova.network.neutron [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.638691] env[62558]: DEBUG nova.compute.manager [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Received event network-vif-plugged-ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.638949] env[62558]: DEBUG oslo_concurrency.lockutils [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] Acquiring lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.642017] env[62558]: DEBUG oslo_concurrency.lockutils [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.642017] env[62558]: DEBUG oslo_concurrency.lockutils [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.642017] env[62558]: DEBUG nova.compute.manager [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] No waiting events found dispatching network-vif-plugged-ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.642017] env[62558]: WARNING nova.compute.manager [req-948dfcec-4ea8-4cc7-9ecf-dbbfdf8fe59a req-9fca7c9f-d4b3-4ec4-8ecc-84f93089a663 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Received unexpected event network-vif-plugged-ec1da108-b883-47de-80b8-b5514917cc4a for instance with vm_state building and task_state spawning. [ 871.647499] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266867, 'name': Rename_Task, 'duration_secs': 0.146295} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.647775] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.648049] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f02f5f2-b665-4281-bff8-a21523c6ad49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.652769] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820c1212-ad8b-4d66-8849-ea01787535f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.660102] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6755fcae-0f09-4672-be39-0d8d4824ec4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.665154] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 871.665154] env[62558]: value = "task-1266868" [ 871.665154] env[62558]: _type = "Task" [ 871.665154] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.696339] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066236e7-f8dd-4137-8952-68ed5b1ea921 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.702470] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266868, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.707308] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595e006f-51cf-405e-a50f-88c6efcf55b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.721990] env[62558]: DEBUG nova.compute.provider_tree [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.739581] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Successfully updated port: ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.846471] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5213f423-579b-edb0-be83-54baef25aa16, 'name': SearchDatastore_Task, 'duration_secs': 0.019852} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.848482] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dfec339-a25d-452d-a202-8e8ba1b9223a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.852886] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 871.852886] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d3789-64d2-4fd5-ca36-b86c4bf2392c" [ 871.852886] env[62558]: _type = "Task" [ 871.852886] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.861272] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d3789-64d2-4fd5-ca36-b86c4bf2392c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.899495] env[62558]: DEBUG nova.objects.instance [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid 4526df2a-f801-4de8-8218-497a7d22034f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.927789] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.177045] env[62558]: DEBUG oslo_vmware.api [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266868, 'name': PowerOnVM_Task, 'duration_secs': 0.515433} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.177507] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.177771] env[62558]: INFO nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Took 8.31 seconds to spawn the instance on the hypervisor. [ 872.178051] env[62558]: DEBUG nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.179934] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1b1384-d372-4c72-91ea-e17690985322 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.225794] env[62558]: DEBUG nova.scheduler.client.report [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.244417] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.244417] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.244417] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.366062] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528d3789-64d2-4fd5-ca36-b86c4bf2392c, 'name': SearchDatastore_Task, 'duration_secs': 0.011157} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.366062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.366062] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/4815ba3f-265f-466a-9850-4c325cdb88de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.366062] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a745ef7-6f60-473c-a38a-d77b1add06e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.373314] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 872.373314] env[62558]: value = "task-1266869" [ 872.373314] env[62558]: _type = "Task" [ 872.373314] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.383634] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.402559] env[62558]: DEBUG nova.objects.base [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance<4526df2a-f801-4de8-8218-497a7d22034f> lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 872.403114] env[62558]: DEBUG nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.469808] env[62558]: DEBUG nova.policy [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.502124] env[62558]: DEBUG nova.network.neutron [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.705591] env[62558]: INFO nova.compute.manager [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Took 39.03 seconds to build instance. [ 872.731556] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.891s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.734668] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.799s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.735120] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.735195] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 872.735662] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.714s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.737533] env[62558]: INFO nova.compute.claims [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.742393] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c2623b-d49b-455b-84fa-d6224383aebf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.757185] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add09f04-9d42-4db5-ab3d-902691540bff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.777157] env[62558]: INFO nova.scheduler.client.report [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Deleted allocations for instance 66ff3b25-d49c-4129-8b51-5338e75b09b1 [ 872.777976] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9786a2-0c3d-4512-9233-831f20af41b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.789313] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4130d60-addd-43ff-b0dd-7fcff4b57fd2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.827882] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179993MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 872.828050] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.829179] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.844569] env[62558]: DEBUG nova.compute.manager [req-cfa9b8c2-df77-4491-90ac-30212becb9ca req-179c3bce-b524-41dd-b404-5b1540710f02 service nova] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Received event network-vif-deleted-3cdad4ed-33d8-46f8-9093-6904227db761 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.844806] env[62558]: DEBUG nova.compute.manager [req-cfa9b8c2-df77-4491-90ac-30212becb9ca req-179c3bce-b524-41dd-b404-5b1540710f02 service nova] [instance: 574525eb-0535-4664-8449-813c16e4781c] Received event network-vif-deleted-8e94d8d1-be4c-4543-aae7-d521cff11fea {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.885762] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266869, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.008736] env[62558]: INFO nova.compute.manager [-] [instance: 574525eb-0535-4664-8449-813c16e4781c] Took 1.50 seconds to deallocate network for instance. [ 873.024380] env[62558]: DEBUG nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Successfully created port: 383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.068444] env[62558]: DEBUG nova.network.neutron [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Updating instance_info_cache with network_info: [{"id": "ec1da108-b883-47de-80b8-b5514917cc4a", "address": "fa:16:3e:40:f1:fe", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1da108-b8", "ovs_interfaceid": "ec1da108-b883-47de-80b8-b5514917cc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.207997] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ae8964c-7fb9-4579-b090-da32ab01888c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.994s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.287673] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6b72ddf4-5059-40e4-9d79-cd6695035900 tempest-ServersAaction247Test-1863330116 tempest-ServersAaction247Test-1863330116-project-member] Lock "66ff3b25-d49c-4129-8b51-5338e75b09b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.779s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.386040] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547895} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.386383] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/4815ba3f-265f-466a-9850-4c325cdb88de.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.386646] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.387387] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58ef8386-2d18-4401-a42b-e349d3422e42 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.394769] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 873.394769] env[62558]: value = "task-1266870" [ 873.394769] env[62558]: _type = "Task" [ 873.394769] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.403801] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.523407] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.570867] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.571213] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Instance network_info: |[{"id": "ec1da108-b883-47de-80b8-b5514917cc4a", "address": "fa:16:3e:40:f1:fe", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1da108-b8", "ovs_interfaceid": "ec1da108-b883-47de-80b8-b5514917cc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.571644] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:f1:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53ebf5df-5ecb-4a0c-a163-d88165639de0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec1da108-b883-47de-80b8-b5514917cc4a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.579265] env[62558]: DEBUG oslo.service.loopingcall [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.579521] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.579768] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b760c77-356f-4e5c-95fe-261f7ac39cbc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.599799] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.599799] env[62558]: value = "task-1266871" [ 873.599799] env[62558]: _type = "Task" [ 873.599799] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.608416] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266871, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.910018] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095287} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.910372] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.912076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a057ea-e7a5-4c9a-ab36-83352755835c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.938202] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/4815ba3f-265f-466a-9850-4c325cdb88de.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.941576] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b74c1151-6b69-4afe-8905-bdeec0e1773c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.958558] env[62558]: DEBUG nova.compute.manager [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Received event network-changed-ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.958751] env[62558]: DEBUG nova.compute.manager [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Refreshing instance network info cache due to event network-changed-ec1da108-b883-47de-80b8-b5514917cc4a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.959060] env[62558]: DEBUG oslo_concurrency.lockutils [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] Acquiring lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.959222] env[62558]: DEBUG oslo_concurrency.lockutils [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] Acquired lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.959377] env[62558]: DEBUG nova.network.neutron [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Refreshing network info cache for port ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.966487] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 873.966487] env[62558]: value = "task-1266872" [ 873.966487] env[62558]: _type = "Task" [ 873.966487] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.977880] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266872, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.080695] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acaacf59-d078-4a4c-b1e0-ebb563c6995e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.089041] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839dbf40-ee88-441c-9bf1-6cc320627aa0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.124834] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc661560-6e83-4781-8982-f1bb6a42e82f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.132932] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266871, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.135962] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b65ed5f-7b8a-4eb6-879b-8b74c7d245a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.150257] env[62558]: DEBUG nova.compute.provider_tree [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.480136] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266872, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.631310] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266871, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.653364] env[62558]: DEBUG nova.scheduler.client.report [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.677349] env[62558]: DEBUG nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Successfully updated port: 383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.895361] env[62558]: DEBUG nova.network.neutron [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Updated VIF entry in instance network info cache for port ec1da108-b883-47de-80b8-b5514917cc4a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.895852] env[62558]: DEBUG nova.network.neutron [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Updating instance_info_cache with network_info: [{"id": "ec1da108-b883-47de-80b8-b5514917cc4a", "address": "fa:16:3e:40:f1:fe", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1da108-b8", "ovs_interfaceid": "ec1da108-b883-47de-80b8-b5514917cc4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.978032] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266872, 'name': ReconfigVM_Task, 'duration_secs': 0.522562} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.978576] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/4815ba3f-265f-466a-9850-4c325cdb88de.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.979554] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec1b76f9-c40b-4098-9dd0-a4b961a1a487 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.986617] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 874.986617] env[62558]: value = "task-1266873" [ 874.986617] env[62558]: _type = "Task" [ 874.986617] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.995576] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266873, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.131908] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266871, 'name': CreateVM_Task, 'duration_secs': 1.485853} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.132136] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.132800] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.132960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.133319] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.133564] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-671eb518-9875-4eab-961e-6854017c2036 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.141429] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 875.141429] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a592ff-b3af-7d44-8b59-e5c7ba8f6d5d" [ 875.141429] env[62558]: _type = "Task" [ 875.141429] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.150793] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a592ff-b3af-7d44-8b59-e5c7ba8f6d5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.168252] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.168833] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.172164] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.728s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.172715] env[62558]: DEBUG nova.objects.instance [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lazy-loading 'resources' on Instance uuid 4a3f7642-1b9c-40d8-973b-5153b559bda0 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.182384] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.182608] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.182806] env[62558]: DEBUG nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.398361] env[62558]: DEBUG oslo_concurrency.lockutils [req-4609427b-314e-4ff5-8282-66bc3e3eba91 req-53f1833f-f85f-4caa-8732-fbaa9c070df5 service nova] Releasing lock "refresh_cache-0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.500327] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266873, 'name': Rename_Task, 'duration_secs': 0.240404} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.500880] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.501424] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-432c1d3c-5fd1-4e26-bd9f-cb71ec2191e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.509687] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 875.509687] env[62558]: value = "task-1266874" [ 875.509687] env[62558]: _type = "Task" [ 875.509687] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.519142] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266874, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.565030] env[62558]: DEBUG nova.compute.manager [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-vif-plugged-383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.565302] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.565513] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.565681] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.565850] env[62558]: DEBUG nova.compute.manager [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] No waiting events found dispatching network-vif-plugged-383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.566031] env[62558]: WARNING nova.compute.manager [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received unexpected event network-vif-plugged-383e1975-fdd7-454f-8cd4-0bce9d1641a9 for instance with vm_state active and task_state None. [ 875.566203] env[62558]: DEBUG nova.compute.manager [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-changed-383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.566361] env[62558]: DEBUG nova.compute.manager [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing instance network info cache due to event network-changed-383e1975-fdd7-454f-8cd4-0bce9d1641a9. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.566530] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.653627] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a592ff-b3af-7d44-8b59-e5c7ba8f6d5d, 'name': SearchDatastore_Task, 'duration_secs': 0.018046} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.654194] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.654591] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.655038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.655482] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.655826] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.656250] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b6dab5e-8f38-4f07-83ca-1ce4b10fac6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.666053] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.666259] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.666992] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9c6bfa8-fe5f-4a8d-bfb0-8aefe573b5d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.672910] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 875.672910] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52815f45-c2a5-df4a-124e-49efdbc2669a" [ 875.672910] env[62558]: _type = "Task" [ 875.672910] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.677138] env[62558]: DEBUG nova.compute.utils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.681141] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.681337] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.692085] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52815f45-c2a5-df4a-124e-49efdbc2669a, 'name': SearchDatastore_Task, 'duration_secs': 0.013526} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.694172] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0310383a-09b2-48ac-aeb2-103dc44d3cb7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.700045] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 875.700045] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52323f7b-297f-b908-0f95-af9b15986f77" [ 875.700045] env[62558]: _type = "Task" [ 875.700045] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.712515] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52323f7b-297f-b908-0f95-af9b15986f77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.742892] env[62558]: WARNING nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 875.921603] env[62558]: DEBUG nova.policy [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61cb5cd30d434f1eb594dd4521dab2d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e1d3e1dcd64e918c1e436e691e290b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.001263] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3f61ad-f873-433d-bb8b-c86d57a8f4b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.009973] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6040be72-0deb-4d25-be76-c54d3fe9528c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.023462] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266874, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.053167] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9128c8b-3592-40ef-ac70-82ae7c9cb300 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.061603] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae97355-d910-49a3-8724-4a0cc189490c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.077264] env[62558]: DEBUG nova.compute.provider_tree [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.184836] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.202489] env[62558]: DEBUG nova.network.neutron [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "383e1975-fdd7-454f-8cd4-0bce9d1641a9", "address": "fa:16:3e:20:51:60", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap383e1975-fd", "ovs_interfaceid": "383e1975-fdd7-454f-8cd4-0bce9d1641a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.214986] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52323f7b-297f-b908-0f95-af9b15986f77, 'name': SearchDatastore_Task, 'duration_secs': 0.012511} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.214986] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.214986] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b/0aeb55df-48e8-4314-ab47-05ff0b5eaa7b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.214986] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-955f5a7a-b38b-4ce5-990f-caaa05b73f01 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.222229] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 876.222229] env[62558]: value = "task-1266875" [ 876.222229] env[62558]: _type = "Task" [ 876.222229] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.232397] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.526406] env[62558]: DEBUG oslo_vmware.api [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1266874, 'name': PowerOnVM_Task, 'duration_secs': 0.590362} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.527898] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.528701] env[62558]: INFO nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Took 10.18 seconds to spawn the instance on the hypervisor. [ 876.534271] env[62558]: DEBUG nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.535174] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fad9314-85a3-44a2-9b67-769de7e19fd3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.585614] env[62558]: DEBUG nova.scheduler.client.report [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.672177] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Successfully created port: 94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.708742] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.709459] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.709675] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.709891] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.710581] env[62558]: DEBUG nova.network.neutron [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Refreshing network info cache for port 383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.711835] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520e769e-5488-46b5-bd38-330162f7cb71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.737019] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.737518] env[62558]: DEBUG nova.virt.hardware [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.746235] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfiguring VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 876.747905] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f822ffd-649c-4328-8483-6e417032fa7e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.767601] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266875, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.769179] env[62558]: DEBUG oslo_vmware.api [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 876.769179] env[62558]: value = "task-1266876" [ 876.769179] env[62558]: _type = "Task" [ 876.769179] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.781572] env[62558]: DEBUG oslo_vmware.api [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266876, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.060545] env[62558]: DEBUG nova.compute.manager [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 877.069734] env[62558]: INFO nova.compute.manager [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Took 41.52 seconds to build instance. [ 877.089562] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.092475] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.990s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.092785] env[62558]: DEBUG nova.objects.instance [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'resources' on Instance uuid 96361b52-b939-4409-aaf7-ba5b3bf054e7 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.125363] env[62558]: INFO nova.scheduler.client.report [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted allocations for instance 4a3f7642-1b9c-40d8-973b-5153b559bda0 [ 877.195803] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.233294] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.233622] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.233780] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.234013] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.235491] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.235679] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.235927] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.236583] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.237027] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.237492] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.237492] env[62558]: DEBUG nova.virt.hardware [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.238778] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fd0f68-28c2-4364-8a29-d6763ffdb1a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.256139] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9650542d-c291-4c86-a49d-095e04a2ae44 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.260470] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.983718} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.262848] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b/0aeb55df-48e8-4314-ab47-05ff0b5eaa7b.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.262848] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.262848] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86d7f4be-b54c-4f46-8d9e-144601c20084 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.284035] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 877.284035] env[62558]: value = "task-1266877" [ 877.284035] env[62558]: _type = "Task" [ 877.284035] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.294324] env[62558]: DEBUG oslo_vmware.api [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.300454] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.574172] env[62558]: DEBUG nova.network.neutron [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updated VIF entry in instance network info cache for port 383e1975-fdd7-454f-8cd4-0bce9d1641a9. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.574625] env[62558]: DEBUG nova.network.neutron [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "383e1975-fdd7-454f-8cd4-0bce9d1641a9", "address": "fa:16:3e:20:51:60", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap383e1975-fd", "ovs_interfaceid": "383e1975-fdd7-454f-8cd4-0bce9d1641a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.577459] env[62558]: DEBUG oslo_concurrency.lockutils [None req-522db5f1-b0a2-4e69-a712-447fe3724f43 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.106s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.591585] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.595387] env[62558]: DEBUG nova.objects.instance [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'numa_topology' on Instance uuid 96361b52-b939-4409-aaf7-ba5b3bf054e7 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.640740] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fea7ba33-fb1c-4043-9b68-dbb7664cd711 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "4a3f7642-1b9c-40d8-973b-5153b559bda0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.744s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.791643] env[62558]: DEBUG oslo_vmware.api [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266876, 'name': ReconfigVM_Task, 'duration_secs': 0.66624} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.792657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.792939] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfigured VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 877.801113] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195361} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.801113] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.801113] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba21aa14-6dd4-44c8-9a4e-f01a632fbe42 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.824598] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b/0aeb55df-48e8-4314-ab47-05ff0b5eaa7b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.825433] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d673d0b-47bf-4f46-946e-e364dc354dec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.845749] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 877.845749] env[62558]: value = "task-1266878" [ 877.845749] env[62558]: _type = "Task" [ 877.845749] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.854529] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.079408] env[62558]: DEBUG oslo_concurrency.lockutils [req-8bfbb5c7-4c80-4c05-b845-18e71b51d4ba req-138157a9-e855-4182-9d22-67963bf2d2cd service nova] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.098346] env[62558]: DEBUG nova.objects.base [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Object Instance<96361b52-b939-4409-aaf7-ba5b3bf054e7> lazy-loaded attributes: resources,numa_topology {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 878.300385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d27f1a00-c20f-409c-96f3-f45fb1a9d05b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.031s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.362297] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.469183] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429be0f7-582f-4cf8-84f2-921ba6badc06 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.478967] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5288f992-c9c5-42fa-904e-67b26f29183d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.518584] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04411dd5-0858-4695-ad8c-bca4642aec98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.528395] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc927660-4cc0-49a1-adf5-9d2a52deb27a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.331174] env[62558]: DEBUG nova.compute.provider_tree [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.332818] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Successfully updated port: 94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.342616] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266878, 'name': ReconfigVM_Task, 'duration_secs': 0.636473} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.343748] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b/0aeb55df-48e8-4314-ab47-05ff0b5eaa7b.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.343748] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-311a0c73-7574-4e12-a231-e819d54b0e21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.350705] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 879.350705] env[62558]: value = "task-1266879" [ 879.350705] env[62558]: _type = "Task" [ 879.350705] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.360411] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266879, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.423448] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 879.424534] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502e3a15-e9c3-4c95-80dd-d15c904f38dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.432152] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 879.432351] env[62558]: ERROR oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk due to incomplete transfer. [ 879.432596] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-23f171e0-64c5-412c-9664-c97992c5aa81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.440323] env[62558]: DEBUG oslo_vmware.rw_handles [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b1758d-0fea-7c79-c62a-6dc4e374202c/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 879.441128] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Uploaded image 2f0f0be9-1e05-4998-a402-10590e152a59 to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 879.445336] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 879.445336] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-18ac5105-7ecb-4327-9dd6-a37821cda532 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.453204] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 879.453204] env[62558]: value = "task-1266880" [ 879.453204] env[62558]: _type = "Task" [ 879.453204] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.465496] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266880, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.654638] env[62558]: DEBUG nova.compute.manager [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Received event network-vif-plugged-94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.654915] env[62558]: DEBUG oslo_concurrency.lockutils [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] Acquiring lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.655216] env[62558]: DEBUG oslo_concurrency.lockutils [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.655398] env[62558]: DEBUG oslo_concurrency.lockutils [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.655571] env[62558]: DEBUG nova.compute.manager [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] No waiting events found dispatching network-vif-plugged-94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.655737] env[62558]: WARNING nova.compute.manager [req-8ca63f29-1e1e-40ef-8d00-df5eaf0f41a6 req-b775a400-555d-4f4c-9544-8f4d4208ae62 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Received unexpected event network-vif-plugged-94ed9b4f-2b00-4433-8c7a-45cfc47135fc for instance with vm_state building and task_state spawning. [ 879.838799] env[62558]: DEBUG nova.scheduler.client.report [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.843062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.843239] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.843545] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.863624] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266879, 'name': Rename_Task, 'duration_secs': 0.252148} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.864901] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 879.865185] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bb48a9f-b0eb-41b4-9749-f8c8f5df14c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.872538] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 879.872538] env[62558]: value = "task-1266881" [ 879.872538] env[62558]: _type = "Task" [ 879.872538] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.880783] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.966623] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266880, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.050579] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.050902] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.350127] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.258s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.357148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.166s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.357937] env[62558]: INFO nova.compute.claims [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.383497] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266881, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.411930] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.469717] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266880, 'name': Destroy_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.554507] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.684895] env[62558]: DEBUG nova.network.neutron [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Updating instance_info_cache with network_info: [{"id": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "address": "fa:16:3e:12:97:68", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ed9b4f-2b", "ovs_interfaceid": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.863619] env[62558]: DEBUG oslo_concurrency.lockutils [None req-919d900f-50af-40fc-b271-d1073f6d881b tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.043s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.867177] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 19.298s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.867420] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.867629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.867799] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.869604] env[62558]: INFO nova.compute.manager [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Terminating instance [ 880.883782] env[62558]: DEBUG nova.compute.manager [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.884456] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.887017] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b29b9404-25db-4a3d-9a6e-e46b7821ec64 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.897223] env[62558]: DEBUG oslo_vmware.api [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266881, 'name': PowerOnVM_Task, 'duration_secs': 0.593056} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.897223] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.897223] env[62558]: INFO nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Took 10.02 seconds to spawn the instance on the hypervisor. [ 880.897223] env[62558]: DEBUG nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.897223] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94694473-29bf-4f6c-96ce-737ad2dfa5a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.901962] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f86889-80d5-4b13-9bdf-eca6b7875875 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.938428] env[62558]: WARNING nova.virt.vmwareapi.vmops [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96361b52-b939-4409-aaf7-ba5b3bf054e7 could not be found. [ 880.938804] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.939100] env[62558]: INFO nova.compute.manager [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 880.939473] env[62558]: DEBUG oslo.service.loopingcall [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.939799] env[62558]: DEBUG nova.compute.manager [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.939958] env[62558]: DEBUG nova.network.neutron [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 880.964201] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266880, 'name': Destroy_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.078427] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.131731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-383e1975-fdd7-454f-8cd4-0bce9d1641a9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.132082] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-383e1975-fdd7-454f-8cd4-0bce9d1641a9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.187625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.187963] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Instance network_info: |[{"id": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "address": "fa:16:3e:12:97:68", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ed9b4f-2b", "ovs_interfaceid": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.188405] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:97:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94ed9b4f-2b00-4433-8c7a-45cfc47135fc', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.203362] env[62558]: DEBUG oslo.service.loopingcall [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.204126] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.204482] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccadd989-fcfc-45fb-b289-dc84457d0233 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.242239] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.242239] env[62558]: value = "task-1266882" [ 881.242239] env[62558]: _type = "Task" [ 881.242239] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.258063] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266882, 'name': CreateVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.430492] env[62558]: INFO nova.compute.manager [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Took 39.72 seconds to build instance. [ 881.473022] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266880, 'name': Destroy_Task, 'duration_secs': 1.947764} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.473022] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Destroyed the VM [ 881.473022] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 881.473022] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1655e5e0-346e-485d-928a-918e827348c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.479516] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 881.479516] env[62558]: value = "task-1266883" [ 881.479516] env[62558]: _type = "Task" [ 881.479516] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.486983] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266883, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.635345] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.635532] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.636558] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee4138d-21e1-4d98-aac3-574568b2a825 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.660740] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489156cf-6997-44f1-8ade-92ef8d968ffe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.663719] env[62558]: DEBUG nova.network.neutron [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.699249] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfiguring VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 881.703686] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a34b481-1223-4335-9206-0abee7f4abff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.717783] env[62558]: DEBUG nova.compute.manager [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Received event network-changed-94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.718035] env[62558]: DEBUG nova.compute.manager [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Refreshing instance network info cache due to event network-changed-94ed9b4f-2b00-4433-8c7a-45cfc47135fc. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.718326] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Acquiring lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.718513] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Acquired lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.718713] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Refreshing network info cache for port 94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.726589] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 881.726589] env[62558]: value = "task-1266884" [ 881.726589] env[62558]: _type = "Task" [ 881.726589] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.731457] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301efb70-7b4d-4c7e-84a5-57a10dea0561 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.739020] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.741894] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af4ae18-a21b-4a98-be7f-6c9911e309a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.781315] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61792d8-cca3-4e59-9eb9-210008284730 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.784076] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266882, 'name': CreateVM_Task, 'duration_secs': 0.494538} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.784248] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.785297] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.785467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.785789] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.786156] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a4a8d1b-f91d-4f11-8965-8048b9e2d4ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.792419] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92660ec9-68fa-4018-a6f3-89176564f35d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.797374] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 881.797374] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527fd44b-55a2-f9cf-3f63-f204d4d47ea6" [ 881.797374] env[62558]: _type = "Task" [ 881.797374] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.809169] env[62558]: DEBUG nova.compute.provider_tree [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.815269] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527fd44b-55a2-f9cf-3f63-f204d4d47ea6, 'name': SearchDatastore_Task, 'duration_secs': 0.010926} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.817518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.817518] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.817518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.817518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.817518] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.817518] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d329b0dc-d48d-45d4-8b25-efe193910f73 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.823927] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.824123] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.824795] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65514cc2-26c2-4782-90b4-e36a54dec3c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.829875] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 881.829875] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3195f-842e-343a-8e90-8e11d6b45b3b" [ 881.829875] env[62558]: _type = "Task" [ 881.829875] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.837258] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3195f-842e-343a-8e90-8e11d6b45b3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.935130] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ebbb1d-606b-49f7-b9b8-c36a6db5f728 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.409s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.999054] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266883, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.172211] env[62558]: INFO nova.compute.manager [-] [instance: 96361b52-b939-4409-aaf7-ba5b3bf054e7] Took 1.23 seconds to deallocate network for instance. [ 882.237212] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.312140] env[62558]: DEBUG nova.scheduler.client.report [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.347656] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c3195f-842e-343a-8e90-8e11d6b45b3b, 'name': SearchDatastore_Task, 'duration_secs': 0.008653} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.348905] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a170a7eb-51c8-40ab-9f90-fd2527d12f6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.356638] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 882.356638] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524752eb-0843-d5bd-7cfa-ed15687d6bc2" [ 882.356638] env[62558]: _type = "Task" [ 882.356638] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.367510] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524752eb-0843-d5bd-7cfa-ed15687d6bc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.439080] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.439399] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.439570] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.439758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.439927] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.442300] env[62558]: INFO nova.compute.manager [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Terminating instance [ 882.444662] env[62558]: DEBUG nova.compute.manager [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.445077] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.445840] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cc756e-0af9-4c03-a09d-e2db7d224053 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.454247] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.454514] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c7a4570-ef0a-470c-8b4c-73c58ed6fb75 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.461590] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 882.461590] env[62558]: value = "task-1266885" [ 882.461590] env[62558]: _type = "Task" [ 882.461590] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.475687] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.494869] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266883, 'name': RemoveSnapshot_Task, 'duration_secs': 0.62242} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.500605] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 882.500605] env[62558]: DEBUG nova.compute.manager [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.500605] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109e9337-63a2-4476-8c13-12d172281424 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.553224] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Updated VIF entry in instance network info cache for port 94ed9b4f-2b00-4433-8c7a-45cfc47135fc. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.553598] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Updating instance_info_cache with network_info: [{"id": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "address": "fa:16:3e:12:97:68", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ed9b4f-2b", "ovs_interfaceid": "94ed9b4f-2b00-4433-8c7a-45cfc47135fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.736981] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.818549] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.819211] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.822649] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.943s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.824190] env[62558]: INFO nova.compute.claims [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.868481] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524752eb-0843-d5bd-7cfa-ed15687d6bc2, 'name': SearchDatastore_Task, 'duration_secs': 0.041054} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.868750] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.869015] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 0e5c1138-7f09-4b76-9c56-93c0ad947ad8/0e5c1138-7f09-4b76-9c56-93c0ad947ad8.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.869285] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0203ab63-0bd2-4fd8-ba27-57e46c9f21a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.876478] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 882.876478] env[62558]: value = "task-1266886" [ 882.876478] env[62558]: _type = "Task" [ 882.876478] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.884737] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.976257] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266885, 'name': PowerOffVM_Task, 'duration_secs': 0.322911} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.976603] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.976832] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.977174] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea3c31c4-309f-4c45-ba83-caf5ef766d11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.011744] env[62558]: INFO nova.compute.manager [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Shelve offloading [ 883.013814] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.014252] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39a2d352-ba55-4310-b1a9-7bcf48ed89b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.020384] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 883.020384] env[62558]: value = "task-1266888" [ 883.020384] env[62558]: _type = "Task" [ 883.020384] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.029200] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 883.029487] env[62558]: DEBUG nova.compute.manager [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.030264] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c5a010-8046-420f-bd5b-cc6d641a6138 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.036601] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.036790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.036963] env[62558]: DEBUG nova.network.neutron [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.039179] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.039381] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.040032] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleting the datastore file [datastore2] 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.040032] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8708638f-f977-4f30-b851-07959195267c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.046707] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 883.046707] env[62558]: value = "task-1266889" [ 883.046707] env[62558]: _type = "Task" [ 883.046707] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.058544] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Releasing lock "refresh_cache-0e5c1138-7f09-4b76-9c56-93c0ad947ad8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.058661] env[62558]: DEBUG nova.compute.manager [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.058896] env[62558]: DEBUG nova.compute.manager [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing instance network info cache due to event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.059179] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.059257] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.059416] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.061104] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.199635] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb3145a5-fbe8-4f99-b53f-a9a9264d7e18 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "96361b52-b939-4409-aaf7-ba5b3bf054e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.332s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.239656] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.330611] env[62558]: DEBUG nova.compute.utils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.333877] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.341621] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.390569] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266886, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.442195] env[62558]: DEBUG nova.policy [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e6e2f1bae4d4bb68070770dbbbe5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78b69738b90142a4943069006e349445', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.555963] env[62558]: DEBUG oslo_vmware.api [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1266889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273062} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.556473] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.556786] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.557298] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.557609] env[62558]: INFO nova.compute.manager [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 883.557963] env[62558]: DEBUG oslo.service.loopingcall [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.558299] env[62558]: DEBUG nova.compute.manager [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.558500] env[62558]: DEBUG nova.network.neutron [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.741658] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.834379] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.892644] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627196} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.893595] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 0e5c1138-7f09-4b76-9c56-93c0ad947ad8/0e5c1138-7f09-4b76-9c56-93c0ad947ad8.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.894069] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.894233] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e59f0e7-a6b6-4b89-af7a-2c86b1a57c04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.901008] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 883.901008] env[62558]: value = "task-1266890" [ 883.901008] env[62558]: _type = "Task" [ 883.901008] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.912530] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.063919] env[62558]: DEBUG nova.network.neutron [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.133751] env[62558]: DEBUG nova.compute.manager [req-5fdee418-ff87-470b-9164-3894c085ccd0 req-91915bd2-148c-4084-89a4-8f3497e04009 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Received event network-vif-deleted-ec1da108-b883-47de-80b8-b5514917cc4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.133751] env[62558]: INFO nova.compute.manager [req-5fdee418-ff87-470b-9164-3894c085ccd0 req-91915bd2-148c-4084-89a4-8f3497e04009 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Neutron deleted interface ec1da108-b883-47de-80b8-b5514917cc4a; detaching it from the instance and deleting it from the info cache [ 884.133751] env[62558]: DEBUG nova.network.neutron [req-5fdee418-ff87-470b-9164-3894c085ccd0 req-91915bd2-148c-4084-89a4-8f3497e04009 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.163254] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updated VIF entry in instance network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.163601] env[62558]: DEBUG nova.network.neutron [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.168718] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Successfully created port: 9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.201911] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee4984b-e8cc-4940-8c11-ba7db99870d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.210523] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1059a5b-473e-4cb8-805c-7c24bba3e7a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.250240] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb913f8f-9eba-4b98-bda2-6376df7ca290 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.262692] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a41a3f-05f7-4426-bd66-8679000eb702 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.266640] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.277556] env[62558]: DEBUG nova.compute.provider_tree [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.352064] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.352149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.415032] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074139} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.415032] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.415664] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f132a27-eead-418d-ad1e-7a53fedd4d0c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.443317] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 0e5c1138-7f09-4b76-9c56-93c0ad947ad8/0e5c1138-7f09-4b76-9c56-93c0ad947ad8.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.443820] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8439dd00-6f88-482a-96a0-0a9aad1af541 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.464833] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 884.464833] env[62558]: value = "task-1266891" [ 884.464833] env[62558]: _type = "Task" [ 884.464833] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.479053] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266891, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.566964] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.609534] env[62558]: DEBUG nova.network.neutron [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.635695] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02aed587-7324-49bc-bf1f-3b6a409ad2f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.646493] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef53d4e-ac6a-47c0-aec9-bf1460561b7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.668927] env[62558]: DEBUG oslo_concurrency.lockutils [req-41f6ee54-6846-4d5a-9833-64b7c7b64dae req-c718c159-d282-4d3d-a041-122617ed292a service nova] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.684446] env[62558]: DEBUG nova.compute.manager [req-5fdee418-ff87-470b-9164-3894c085ccd0 req-91915bd2-148c-4084-89a4-8f3497e04009 service nova] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Detach interface failed, port_id=ec1da108-b883-47de-80b8-b5514917cc4a, reason: Instance 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 884.756507] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.780430] env[62558]: DEBUG nova.scheduler.client.report [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.832324] env[62558]: DEBUG nova.compute.manager [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-vif-unplugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.832661] env[62558]: DEBUG oslo_concurrency.lockutils [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.832882] env[62558]: DEBUG oslo_concurrency.lockutils [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.833077] env[62558]: DEBUG oslo_concurrency.lockutils [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.833262] env[62558]: DEBUG nova.compute.manager [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] No waiting events found dispatching network-vif-unplugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.833437] env[62558]: WARNING nova.compute.manager [req-52532b0b-4815-4720-99cd-16e645bee144 req-417173b7-80c3-4480-9721-0be2b9dbd7bc service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received unexpected event network-vif-unplugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b for instance with vm_state shelved and task_state shelving_offloading. [ 884.846778] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.854550] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.884246] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.884565] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.884736] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.884926] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.885269] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.885472] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.885695] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.885861] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.886228] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.886462] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.886810] env[62558]: DEBUG nova.virt.hardware [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.888421] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff6299c-9392-47f2-a8cc-f0e6ab7ce7ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.899245] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b195dc62-4336-486d-8547-142f280b9f1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.978130] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266891, 'name': ReconfigVM_Task, 'duration_secs': 0.27241} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.978130] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 0e5c1138-7f09-4b76-9c56-93c0ad947ad8/0e5c1138-7f09-4b76-9c56-93c0ad947ad8.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.978478] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4d22cc8-37eb-4bf1-bf10-95d25eca82cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.985632] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 884.985632] env[62558]: value = "task-1266892" [ 884.985632] env[62558]: _type = "Task" [ 884.985632] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.997113] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266892, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.029476] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.030768] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3baedf33-bfe6-40ea-8bc2-4759114c0dd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.038995] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.039301] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f799830-e334-4a27-8eef-02fa0cf5e9a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.112803] env[62558]: INFO nova.compute.manager [-] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Took 1.55 seconds to deallocate network for instance. [ 885.152750] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.153017] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.153219] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleting the datastore file [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.153541] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5bfbd14-0848-4747-8804-b2cf65044ccc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.161429] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 885.161429] env[62558]: value = "task-1266894" [ 885.161429] env[62558]: _type = "Task" [ 885.161429] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.178135] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.257041] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.291021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.291021] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.292392] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.292752] env[62558]: DEBUG nova.objects.instance [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lazy-loading 'resources' on Instance uuid 6243631f-b240-4d7c-8910-0bf3d2dedb77 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.384283] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.496914] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266892, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.621789] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.675602] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.763237] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.796046] env[62558]: DEBUG nova.compute.utils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.802319] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.802319] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.896683] env[62558]: DEBUG nova.policy [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.004177] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266892, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.177385] env[62558]: DEBUG oslo_vmware.api [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1266894, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.555682} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.177636] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.177825] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.178011] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.188071] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336827cf-c930-49cb-8bc0-1787b1d74cc8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.195594] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2addfc-2b82-4a85-80a3-4e695869fd11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.201564] env[62558]: INFO nova.scheduler.client.report [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted allocations for instance 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 [ 886.233724] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5504df6f-21bc-42f1-ab76-996dfdde6fe5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.241713] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b7f366-1d52-4fc9-9d80-cc965b1ee665 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.260443] env[62558]: DEBUG nova.compute.provider_tree [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.268833] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.302235] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.390193] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Successfully updated port: 9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.503202] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266892, 'name': Rename_Task, 'duration_secs': 1.145477} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.503520] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.503764] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa4847f0-6e57-47fa-bab4-7d5bd607af8d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.510738] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 886.510738] env[62558]: value = "task-1266895" [ 886.510738] env[62558]: _type = "Task" [ 886.510738] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.520652] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.530206] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Successfully created port: 031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.734788] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.759761] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.765856] env[62558]: DEBUG nova.scheduler.client.report [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.893699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.893900] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.893976] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.916701] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.916916] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing instance network info cache due to event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.919019] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.919019] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.919019] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.020396] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266895, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.260728] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.271689] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.979s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.276679] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.704s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.276679] env[62558]: DEBUG nova.objects.instance [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lazy-loading 'resources' on Instance uuid 80221843-4e15-4f20-aeb4-4e6081371b95 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.296661] env[62558]: INFO nova.scheduler.client.report [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted allocations for instance 6243631f-b240-4d7c-8910-0bf3d2dedb77 [ 887.312599] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.340898] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.341223] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.341436] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.341820] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.341857] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.342637] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.342895] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.343106] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.343293] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.343495] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.343688] env[62558]: DEBUG nova.virt.hardware [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.345041] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffe5a77-4336-4809-be1c-d33dd3d41a3a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.355158] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa256639-7aed-4d43-abe2-7af5e422fbce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.436256] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.525890] env[62558]: DEBUG oslo_vmware.api [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266895, 'name': PowerOnVM_Task, 'duration_secs': 0.653766} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.526194] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.526915] env[62558]: INFO nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Took 10.33 seconds to spawn the instance on the hypervisor. [ 887.526915] env[62558]: DEBUG nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.527360] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305ec12f-8cd4-40c9-8c32-7a0244a09ef2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.688114] env[62558]: DEBUG nova.network.neutron [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Updating instance_info_cache with network_info: [{"id": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "address": "fa:16:3e:9f:bb:b3", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba08c86-31", "ovs_interfaceid": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.715190] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updated VIF entry in instance network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.715561] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7472a75c-48", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.758927] env[62558]: DEBUG oslo_vmware.api [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266884, 'name': ReconfigVM_Task, 'duration_secs': 5.732943} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.759580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.759854] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Reconfigured VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 887.809158] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50044647-af13-4792-b1a7-f84156aa8818 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "6243631f-b240-4d7c-8910-0bf3d2dedb77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.403s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.050420] env[62558]: INFO nova.compute.manager [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Took 38.05 seconds to build instance. [ 888.092505] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ad0c7b-120a-4658-8a12-fd2a9db264ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.101216] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdde2ef8-a7dd-4f35-9a8f-e95eae9fe559 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.138611] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc465365-57a7-42a6-8c08-0819702b820c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.146258] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86f2878-71b9-419a-9bff-565d6c30c0a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.159959] env[62558]: DEBUG nova.compute.provider_tree [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.191490] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.192070] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance network_info: |[{"id": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "address": "fa:16:3e:9f:bb:b3", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba08c86-31", "ovs_interfaceid": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.192564] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:bb:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ba08c86-31c9-4715-a527-8ae92d12af2e', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.204038] env[62558]: DEBUG oslo.service.loopingcall [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.204115] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.204312] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-477cd91f-8491-4b90-a5dc-1562800d98d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.221900] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Successfully updated port: 031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.223370] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.223663] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Received event network-vif-plugged-9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.223787] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.223985] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.224222] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.225086] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] No waiting events found dispatching network-vif-plugged-9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.225086] env[62558]: WARNING nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Received unexpected event network-vif-plugged-9ba08c86-31c9-4715-a527-8ae92d12af2e for instance with vm_state building and task_state spawning. [ 888.225086] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Received event network-changed-9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.225086] env[62558]: DEBUG nova.compute.manager [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Refreshing instance network info cache due to event network-changed-9ba08c86-31c9-4715-a527-8ae92d12af2e. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.225250] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Acquiring lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.225345] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Acquired lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.225641] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Refreshing network info cache for port 9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.229818] env[62558]: DEBUG nova.compute.manager [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-vif-deleted-383e1975-fdd7-454f-8cd4-0bce9d1641a9 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.230018] env[62558]: INFO nova.compute.manager [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Neutron deleted interface 383e1975-fdd7-454f-8cd4-0bce9d1641a9; detaching it from the instance and deleting it from the info cache [ 888.232020] env[62558]: DEBUG nova.network.neutron [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.234998] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.234998] env[62558]: value = "task-1266899" [ 888.234998] env[62558]: _type = "Task" [ 888.234998] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.247277] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266899, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.553618] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e22d185-b67a-4982-b650-31191c712069 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.562s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.663255] env[62558]: DEBUG nova.scheduler.client.report [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.728916] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.729090] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.729251] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.734282] env[62558]: DEBUG oslo_concurrency.lockutils [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.735202] env[62558]: DEBUG oslo_concurrency.lockutils [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] Acquired lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.735354] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d50fa5-4f8c-4b86-b7ce-a7f40242874c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.759025] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266899, 'name': CreateVM_Task, 'duration_secs': 0.404004} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.759374] env[62558]: DEBUG oslo_concurrency.lockutils [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] Releasing lock "4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.759632] env[62558]: WARNING nova.compute.manager [req-deee2093-08fe-4254-ade2-211a06d13423 req-3b9d83a9-de31-4da8-be0d-e6c4ffef6342 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Detach interface failed, port_id=383e1975-fdd7-454f-8cd4-0bce9d1641a9, reason: No device with interface-id 383e1975-fdd7-454f-8cd4-0bce9d1641a9 exists on VM: nova.exception.NotFound: No device with interface-id 383e1975-fdd7-454f-8cd4-0bce9d1641a9 exists on VM [ 888.759942] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.760591] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.760750] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.761104] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.761368] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adca84db-57a1-4b81-814b-7ae91239998f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.768550] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 888.768550] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52289ddb-c57a-d336-a499-12950ced697b" [ 888.768550] env[62558]: _type = "Task" [ 888.768550] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.777293] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52289ddb-c57a-d336-a499-12950ced697b, 'name': SearchDatastore_Task} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.777565] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.777791] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.778033] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.778188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.778385] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.778623] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cb8bc1c-10e7-4d87-850f-721f339a4772 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.785953] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.786148] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.786825] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543d0a63-972d-46e2-9b0f-32ee974db8a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.791741] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 888.791741] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52903d3b-4ce2-1fff-8151-308e2186edc6" [ 888.791741] env[62558]: _type = "Task" [ 888.791741] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.792140] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.801222] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52903d3b-4ce2-1fff-8151-308e2186edc6, 'name': SearchDatastore_Task, 'duration_secs': 0.007157} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.801907] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ae734d-2c4c-4298-a747-7428c147b1f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.806410] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 888.806410] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5291697b-0554-54ae-55e0-0e06b7755332" [ 888.806410] env[62558]: _type = "Task" [ 888.806410] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.813767] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5291697b-0554-54ae-55e0-0e06b7755332, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.086297] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.086581] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.086859] env[62558]: DEBUG nova.network.neutron [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.095670] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Updated VIF entry in instance network info cache for port 9ba08c86-31c9-4715-a527-8ae92d12af2e. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.096081] env[62558]: DEBUG nova.network.neutron [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Updating instance_info_cache with network_info: [{"id": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "address": "fa:16:3e:9f:bb:b3", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ba08c86-31", "ovs_interfaceid": "9ba08c86-31c9-4715-a527-8ae92d12af2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.137364] env[62558]: DEBUG nova.compute.manager [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Received event network-vif-plugged-031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.137364] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Acquiring lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.137364] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.137364] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.137364] env[62558]: DEBUG nova.compute.manager [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] No waiting events found dispatching network-vif-plugged-031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.137364] env[62558]: WARNING nova.compute.manager [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Received unexpected event network-vif-plugged-031eaee8-ff01-4bbf-8d37-5db32153834f for instance with vm_state building and task_state spawning. [ 889.137364] env[62558]: DEBUG nova.compute.manager [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Received event network-changed-031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.137364] env[62558]: DEBUG nova.compute.manager [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Refreshing instance network info cache due to event network-changed-031eaee8-ff01-4bbf-8d37-5db32153834f. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.137364] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Acquiring lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.168527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.173893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.244s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.173893] env[62558]: DEBUG nova.objects.instance [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lazy-loading 'resources' on Instance uuid 2b8430ef-c8eb-4eb8-a754-3c552662b966 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.196785] env[62558]: INFO nova.scheduler.client.report [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Deleted allocations for instance 80221843-4e15-4f20-aeb4-4e6081371b95 [ 889.279646] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.318909] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5291697b-0554-54ae-55e0-0e06b7755332, 'name': SearchDatastore_Task, 'duration_secs': 0.007258} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.319213] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.319473] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.319740] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b0a5336-caa9-48f7-a93c-f08f7614d4a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.327670] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 889.327670] env[62558]: value = "task-1266900" [ 889.327670] env[62558]: _type = "Task" [ 889.327670] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.335483] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.408821] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.409125] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.409375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "4526df2a-f801-4de8-8218-497a7d22034f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.410058] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.410230] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.413112] env[62558]: INFO nova.compute.manager [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Terminating instance [ 889.416095] env[62558]: DEBUG nova.compute.manager [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.416095] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.416302] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d0a43a-135f-471a-b745-34d3ce27afe6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.423804] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.423955] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a689ad22-bf37-4c7e-9fa5-2ea2c21b8e11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.429692] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 889.429692] env[62558]: value = "task-1266901" [ 889.429692] env[62558]: _type = "Task" [ 889.429692] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.440370] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.488743] env[62558]: DEBUG nova.compute.manager [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.490176] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac30f36a-407c-4889-8868-f58dbf76935d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.494937] env[62558]: DEBUG nova.network.neutron [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Updating instance_info_cache with network_info: [{"id": "031eaee8-ff01-4bbf-8d37-5db32153834f", "address": "fa:16:3e:a0:2f:8b", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031eaee8-ff", "ovs_interfaceid": "031eaee8-ff01-4bbf-8d37-5db32153834f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.600190] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7481802-c3bc-4579-a9e3-6d694344e26d req-1cae17fa-c1e1-4b84-8ce5-2097ce56e6b5 service nova] Releasing lock "refresh_cache-6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.708591] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27aabae2-dfa4-46b3-aa10-c8e784e96d6f tempest-ServerTagsTestJSON-1780423778 tempest-ServerTagsTestJSON-1780423778-project-member] Lock "80221843-4e15-4f20-aeb4-4e6081371b95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.102s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.809265] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.809567] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.809797] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.810161] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.810416] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.815555] env[62558]: INFO nova.compute.manager [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Terminating instance [ 889.818278] env[62558]: DEBUG nova.compute.manager [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.818656] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.824025] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541adf6f-3ea7-44af-a395-bf26459c7c8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.835544] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.837025] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac4204f9-e381-47e9-aed6-a79e83b5b61a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.843661] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266900, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.845256] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 889.845256] env[62558]: value = "task-1266903" [ 889.845256] env[62558]: _type = "Task" [ 889.845256] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.860023] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266903, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.942967] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266901, 'name': PowerOffVM_Task, 'duration_secs': 0.192909} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.943212] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.943492] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.943737] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfaa61ec-e55f-4254-bbad-b8188e06fbba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.948904] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.949272] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.949527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.949746] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.949942] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.952411] env[62558]: INFO nova.compute.manager [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Terminating instance [ 889.959272] env[62558]: DEBUG nova.compute.manager [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.959562] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.960487] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918f9c99-37dc-4e33-856e-f91393b8c6a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.978026] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.978026] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a536162-9936-4b98-ba67-84363e5d85bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.985149] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 889.985149] env[62558]: value = "task-1266905" [ 889.985149] env[62558]: _type = "Task" [ 889.985149] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.994811] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.998059] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.998570] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Instance network_info: |[{"id": "031eaee8-ff01-4bbf-8d37-5db32153834f", "address": "fa:16:3e:a0:2f:8b", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031eaee8-ff", "ovs_interfaceid": "031eaee8-ff01-4bbf-8d37-5db32153834f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.999303] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Acquired lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.999655] env[62558]: DEBUG nova.network.neutron [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Refreshing network info cache for port 031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.001257] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:2f:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '031eaee8-ff01-4bbf-8d37-5db32153834f', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.009852] env[62558]: DEBUG oslo.service.loopingcall [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.011792] env[62558]: INFO nova.compute.manager [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] instance snapshotting [ 890.014248] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.019287] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-538ee278-5174-4b97-8af1-de95f037f2d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.038942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c981159-b8f7-4997-a283-ba78668739f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.041893] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.042181] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.042822] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleting the datastore file [datastore1] 4526df2a-f801-4de8-8218-497a7d22034f {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.043390] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e60801a1-17eb-49ac-ac69-3887578258a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.050436] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.050436] env[62558]: value = "task-1266906" [ 890.050436] env[62558]: _type = "Task" [ 890.050436] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.070811] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 890.070811] env[62558]: value = "task-1266907" [ 890.070811] env[62558]: _type = "Task" [ 890.070811] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.075065] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bce0cd-872b-4ab3-89f7-1d53cef28e61 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.084764] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266906, 'name': CreateVM_Task} progress is 15%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.085917] env[62558]: DEBUG nova.network.neutron [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [{"id": "9d645945-9804-4520-bd40-e7475a0dd042", "address": "fa:16:3e:ee:6f:13", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d645945-98", "ovs_interfaceid": "9d645945-9804-4520-bd40-e7475a0dd042", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.097280] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.117248] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93271486-e16e-4531-a88c-645d47c95d77 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.127423] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24031eab-7f61-4dff-908d-6cd5c441ff4a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.166634] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8428651-935b-4a4a-b47b-cd59373b0e15 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.177942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08cde44c-5807-4514-94ad-a89237fcde45 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.198756] env[62558]: DEBUG nova.compute.provider_tree [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.340294] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551112} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.341247] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.341543] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.341822] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c211266-7159-4d7f-a5a7-f7125ad06fb3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.348676] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 890.348676] env[62558]: value = "task-1266908" [ 890.348676] env[62558]: _type = "Task" [ 890.348676] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.362257] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266903, 'name': PowerOffVM_Task, 'duration_secs': 0.215892} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.365389] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.365581] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.365868] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.366102] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f79f41b-7c16-4f88-98d1-b0b47b290dfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.445781] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.446022] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.446212] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleting the datastore file [datastore2] 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.446492] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4fe8719-386e-4212-a2b6-9229385e42f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.454686] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 890.454686] env[62558]: value = "task-1266910" [ 890.454686] env[62558]: _type = "Task" [ 890.454686] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.464643] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266910, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.498574] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.567441] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266906, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.586985] env[62558]: DEBUG oslo_vmware.api [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1266907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.403789} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.587895] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.587895] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.587895] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.588082] env[62558]: INFO nova.compute.manager [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 890.589189] env[62558]: DEBUG oslo.service.loopingcall [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.589189] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-4526df2a-f801-4de8-8218-497a7d22034f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.590902] env[62558]: DEBUG nova.compute.manager [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.591100] env[62558]: DEBUG nova.network.neutron [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.603370] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 890.603370] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-42f02866-2a9c-4891-bc35-c4229022f0d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.609186] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 890.609186] env[62558]: value = "task-1266911" [ 890.609186] env[62558]: _type = "Task" [ 890.609186] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.617663] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266911, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.709021] env[62558]: DEBUG nova.scheduler.client.report [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.841743] env[62558]: DEBUG nova.network.neutron [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Updated VIF entry in instance network info cache for port 031eaee8-ff01-4bbf-8d37-5db32153834f. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.842229] env[62558]: DEBUG nova.network.neutron [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Updating instance_info_cache with network_info: [{"id": "031eaee8-ff01-4bbf-8d37-5db32153834f", "address": "fa:16:3e:a0:2f:8b", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031eaee8-ff", "ovs_interfaceid": "031eaee8-ff01-4bbf-8d37-5db32153834f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.864485] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068498} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.865081] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.865854] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c1873d-93eb-4d11-ad1d-2c59c9b3d8c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.890728] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.891405] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67368e37-95e8-4f35-9344-6ee2be9fa781 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.911796] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 890.911796] env[62558]: value = "task-1266912" [ 890.911796] env[62558]: _type = "Task" [ 890.911796] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.920637] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266912, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.964643] env[62558]: DEBUG oslo_vmware.api [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266910, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147514} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.964868] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.965165] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.965504] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.965809] env[62558]: INFO nova.compute.manager [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Took 1.15 seconds to destroy the instance on the hypervisor. [ 890.966259] env[62558]: DEBUG oslo.service.loopingcall [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.966572] env[62558]: DEBUG nova.compute.manager [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.966683] env[62558]: DEBUG nova.network.neutron [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.995933] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266905, 'name': PowerOffVM_Task, 'duration_secs': 0.737047} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.997488] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.997488] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.997488] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfdbe9e2-1ebc-4e66-a469-6c007aa85a2e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.065892] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266906, 'name': CreateVM_Task, 'duration_secs': 0.577946} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.066072] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.067078] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.067078] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.067299] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.067566] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6149d345-8708-48d0-ae5d-6a6348171f7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.076751] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 891.076751] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526cf860-39df-99a2-c4de-3d78b830d08c" [ 891.076751] env[62558]: _type = "Task" [ 891.076751] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.083704] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.085463] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.085947] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleting the datastore file [datastore2] a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.088974] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-937bbe77-9981-420f-b8e9-f4024d22f915 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.090841] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526cf860-39df-99a2-c4de-3d78b830d08c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.095905] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3f31c32b-9565-4493-9402-565b0fbb926b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-4526df2a-f801-4de8-8218-497a7d22034f-383e1975-fdd7-454f-8cd4-0bce9d1641a9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.964s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.096981] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for the task: (returnval){ [ 891.096981] env[62558]: value = "task-1266914" [ 891.096981] env[62558]: _type = "Task" [ 891.096981] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.108944] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.118725] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266911, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.211241] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.216154] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.387s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.242413] env[62558]: INFO nova.scheduler.client.report [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Deleted allocations for instance 2b8430ef-c8eb-4eb8-a754-3c552662b966 [ 891.346517] env[62558]: DEBUG oslo_concurrency.lockutils [req-fbe84988-2fbd-428c-9cbe-d88161f9e3be req-20396b19-0483-4a13-8008-499bc6af58f3 service nova] Releasing lock "refresh_cache-016c31f6-648c-41f5-909f-5a1ae6366b98" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.372680] env[62558]: DEBUG nova.compute.manager [req-bc806fda-dd4a-4fda-8aec-6cc6705aaabc req-793dbccb-cd1c-41b4-b8dd-fdcb4028c803 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Received event network-vif-deleted-74146c35-c7d1-4654-9100-87b81d3d4ea4 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.372861] env[62558]: INFO nova.compute.manager [req-bc806fda-dd4a-4fda-8aec-6cc6705aaabc req-793dbccb-cd1c-41b4-b8dd-fdcb4028c803 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Neutron deleted interface 74146c35-c7d1-4654-9100-87b81d3d4ea4; detaching it from the instance and deleting it from the info cache [ 891.373632] env[62558]: DEBUG nova.network.neutron [req-bc806fda-dd4a-4fda-8aec-6cc6705aaabc req-793dbccb-cd1c-41b4-b8dd-fdcb4028c803 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.402061] env[62558]: DEBUG nova.compute.manager [req-0f5be823-bd25-4d71-abcb-b1d3db9cf6e8 req-23685b27-59c6-4a73-8dd6-ef2f2e5cb122 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Received event network-vif-deleted-9d645945-9804-4520-bd40-e7475a0dd042 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.402061] env[62558]: INFO nova.compute.manager [req-0f5be823-bd25-4d71-abcb-b1d3db9cf6e8 req-23685b27-59c6-4a73-8dd6-ef2f2e5cb122 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Neutron deleted interface 9d645945-9804-4520-bd40-e7475a0dd042; detaching it from the instance and deleting it from the info cache [ 891.402061] env[62558]: DEBUG nova.network.neutron [req-0f5be823-bd25-4d71-abcb-b1d3db9cf6e8 req-23685b27-59c6-4a73-8dd6-ef2f2e5cb122 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.425146] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.589220] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526cf860-39df-99a2-c4de-3d78b830d08c, 'name': SearchDatastore_Task, 'duration_secs': 0.041617} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.591127] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.591127] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.591127] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.591127] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.591127] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.591127] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-868275ff-5dc2-4021-a45e-b5d13c66b919 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.605021] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.605021] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.605021] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5194ace-af39-4285-8393-ca8a01312512 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.612619] env[62558]: DEBUG oslo_vmware.api [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Task: {'id': task-1266914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.427335} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.616392] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.616889] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.617172] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.617421] env[62558]: INFO nova.compute.manager [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Took 1.66 seconds to destroy the instance on the hypervisor. [ 891.617742] env[62558]: DEBUG oslo.service.loopingcall [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.622016] env[62558]: DEBUG nova.compute.manager [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.622016] env[62558]: DEBUG nova.network.neutron [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 891.622016] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 891.622016] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5268b14b-5504-8f11-1429-1133ae246f26" [ 891.622016] env[62558]: _type = "Task" [ 891.622016] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.629920] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266911, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.637541] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5268b14b-5504-8f11-1429-1133ae246f26, 'name': SearchDatastore_Task, 'duration_secs': 0.012827} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.638469] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca0d8c04-07cb-4b6d-b83e-7d251f2a8b99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.644398] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 891.644398] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e45809-4a57-a172-e269-2d412d508ca8" [ 891.644398] env[62558]: _type = "Task" [ 891.644398] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.654977] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e45809-4a57-a172-e269-2d412d508ca8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.752312] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d88e2671-12c6-4bdb-adae-34dea6c58dd7 tempest-ServersTestMultiNic-1143990576 tempest-ServersTestMultiNic-1143990576-project-member] Lock "2b8430ef-c8eb-4eb8-a754-3c552662b966" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.119s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.845105] env[62558]: DEBUG nova.network.neutron [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.848319] env[62558]: DEBUG nova.network.neutron [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.878363] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-229c8237-7c8f-4f1d-8711-aa1de9999f6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.890846] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5294c5a0-e4ba-4531-91d0-098a91fd2cb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.904726] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39f2e2ef-d30f-4bae-a6ef-abcd6479ce52 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.914786] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed56ec8-538a-4c13-a0b3-c9db1cea478f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.939912] env[62558]: DEBUG nova.compute.manager [req-bc806fda-dd4a-4fda-8aec-6cc6705aaabc req-793dbccb-cd1c-41b4-b8dd-fdcb4028c803 service nova] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Detach interface failed, port_id=74146c35-c7d1-4654-9100-87b81d3d4ea4, reason: Instance 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 891.949689] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266912, 'name': ReconfigVM_Task, 'duration_secs': 0.859084} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.950633] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.951591] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22e63f73-8f8b-4ac7-8993-91d635bdf37d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.963856] env[62558]: DEBUG nova.compute.manager [req-0f5be823-bd25-4d71-abcb-b1d3db9cf6e8 req-23685b27-59c6-4a73-8dd6-ef2f2e5cb122 service nova] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Detach interface failed, port_id=9d645945-9804-4520-bd40-e7475a0dd042, reason: Instance 4526df2a-f801-4de8-8218-497a7d22034f could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 891.969785] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 891.969785] env[62558]: value = "task-1266916" [ 891.969785] env[62558]: _type = "Task" [ 891.969785] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.982641] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266916, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.121812] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266911, 'name': CreateSnapshot_Task, 'duration_secs': 1.16639} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.122152] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 892.122963] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e8ffcb-e8f6-45af-96a1-d5eb5b1ed128 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.160897] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e45809-4a57-a172-e269-2d412d508ca8, 'name': SearchDatastore_Task, 'duration_secs': 0.010552} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.160897] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.161465] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 016c31f6-648c-41f5-909f-5a1ae6366b98/016c31f6-648c-41f5-909f-5a1ae6366b98.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.161674] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63ec921e-9c6a-4004-a042-19d310d8ef37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.170155] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 892.170155] env[62558]: value = "task-1266917" [ 892.170155] env[62558]: _type = "Task" [ 892.170155] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.180528] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.231679] env[62558]: INFO nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating resource usage from migration d5cc95be-a6ee-4246-bcb7-88ff1cd2f615 [ 892.255950] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ac801d7-af70-46e0-88b3-02caee13497d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256144] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256270] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a1d242e6-1561-4bd4-8e39-281ab6346661 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256417] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256506] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4526df2a-f801-4de8-8218-497a7d22034f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256617] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 49a58b46-207f-4515-b313-afcdb2d1ced3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256772] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.256983] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.257240] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 574525eb-0535-4664-8449-813c16e4781c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 892.257449] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4815ba3f-265f-466a-9850-4c325cdb88de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.257666] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 892.257870] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 0e5c1138-7f09-4b76-9c56-93c0ad947ad8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.258068] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.258211] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 016c31f6-648c-41f5-909f-5a1ae6366b98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.258333] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Migration d5cc95be-a6ee-4246-bcb7-88ff1cd2f615 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 892.258454] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 63050323-047e-4d73-91ae-859467b4b5a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 892.348026] env[62558]: INFO nova.compute.manager [-] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Took 1.38 seconds to deallocate network for instance. [ 892.357664] env[62558]: INFO nova.compute.manager [-] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Took 1.77 seconds to deallocate network for instance. [ 892.480044] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266916, 'name': Rename_Task, 'duration_secs': 0.258686} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.480370] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.480633] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4bc95b0-2a6f-4eeb-aade-09449e71125a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.487985] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 892.487985] env[62558]: value = "task-1266918" [ 892.487985] env[62558]: _type = "Task" [ 892.487985] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.495886] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.509404] env[62558]: DEBUG nova.network.neutron [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.644237] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 892.644237] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-96bc48e8-5d36-4075-b747-6aed123cbc2a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.652096] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 892.652096] env[62558]: value = "task-1266919" [ 892.652096] env[62558]: _type = "Task" [ 892.652096] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.667861] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.683792] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266917, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.762469] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 7aebdb8a-517b-4168-91e0-1b704e6a11c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 892.855617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.865135] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.997660] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266918, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.012569] env[62558]: INFO nova.compute.manager [-] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Took 1.39 seconds to deallocate network for instance. [ 893.181538] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.207379] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81477} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.208412] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 016c31f6-648c-41f5-909f-5a1ae6366b98/016c31f6-648c-41f5-909f-5a1ae6366b98.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.208841] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.209382] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3687d0a0-024c-47bb-befd-845602aef241 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.227041] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 893.227041] env[62558]: value = "task-1266920" [ 893.227041] env[62558]: _type = "Task" [ 893.227041] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.242026] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.266403] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ecf01ab6-5019-4453-b102-0e754abc4ef8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 893.266593] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (dc830c09-1c36-446a-8af3-d3826bec8b3b): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 893.266877] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 893.267085] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 893.433565] env[62558]: DEBUG nova.compute.manager [req-70dcc0bd-2f09-4985-ac23-cb6dcc034223 req-e7ce3f97-4947-4b64-a46e-d3da645cb7c5 service nova] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Received event network-vif-deleted-332b0302-a7b5-4574-b8ef-006d49069a3a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.499264] env[62558]: DEBUG oslo_vmware.api [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266918, 'name': PowerOnVM_Task, 'duration_secs': 0.635236} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.500071] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.500122] env[62558]: INFO nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Took 8.65 seconds to spawn the instance on the hypervisor. [ 893.501564] env[62558]: DEBUG nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.501564] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30082852-a981-4c38-82a5-adebb1a6df47 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.519362] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.545678] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e2304a-0e40-4ba0-98bf-d37a0fb7d2e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.553810] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca736408-1fa8-4dd6-9311-49dc8513e090 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.583877] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164141dc-a321-4c23-98ca-4d23065ba3c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.591703] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3215f2d-967d-4d0b-9b28-636d983f3d07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.605192] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.664320] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.737489] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093477} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.737774] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.738561] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13ed10d-74a5-4763-9aee-3409cbb67f37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.761228] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 016c31f6-648c-41f5-909f-5a1ae6366b98/016c31f6-648c-41f5-909f-5a1ae6366b98.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.761500] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1bfb1bb-4c0f-44fe-a19f-a3a8a6062ff9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.782712] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 893.782712] env[62558]: value = "task-1266921" [ 893.782712] env[62558]: _type = "Task" [ 893.782712] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.790840] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.022827] env[62558]: INFO nova.compute.manager [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Took 31.85 seconds to build instance. [ 894.108207] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.164558] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.303065] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266921, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.524823] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6c99555b-fef9-4539-b112-3c11caa7e0ce tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.361s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.614533] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 894.614854] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.400s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.615170] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.092s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.615361] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.617964] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 17.029s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.640339] env[62558]: INFO nova.scheduler.client.report [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Deleted allocations for instance 574525eb-0535-4664-8449-813c16e4781c [ 894.666513] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task} progress is 95%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.789836] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.790139] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.790329] env[62558]: DEBUG nova.compute.manager [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.794659] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24900f52-34eb-49bb-a264-55b17dcc175b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.802609] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266921, 'name': ReconfigVM_Task, 'duration_secs': 0.889099} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.804161] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 016c31f6-648c-41f5-909f-5a1ae6366b98/016c31f6-648c-41f5-909f-5a1ae6366b98.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.804847] env[62558]: DEBUG nova.compute.manager [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 894.805604] env[62558]: DEBUG nova.objects.instance [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'flavor' on Instance uuid 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.808154] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-35d6a5ad-5e28-4249-b967-61d704b98b56 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.813229] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 894.813229] env[62558]: value = "task-1266923" [ 894.813229] env[62558]: _type = "Task" [ 894.813229] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.821540] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266923, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.127077] env[62558]: INFO nova.compute.claims [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.151051] env[62558]: DEBUG oslo_concurrency.lockutils [None req-338ae310-a383-46ce-a7f8-f39197a0e2c2 tempest-ServerGroupTestJSON-1377320768 tempest-ServerGroupTestJSON-1377320768-project-member] Lock "574525eb-0535-4664-8449-813c16e4781c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.800s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.171207] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266919, 'name': CloneVM_Task, 'duration_secs': 2.1977} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.171207] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Created linked-clone VM from snapshot [ 895.172556] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcaa560-4bec-4779-83d8-e564706f5417 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.188111] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Uploading image fe8845c5-445d-4f71-8ce6-a1a0ce770a67 {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 895.208467] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 895.208467] env[62558]: value = "vm-272561" [ 895.208467] env[62558]: _type = "VirtualMachine" [ 895.208467] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 895.209829] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1bf81f14-a775-4b40-9b07-638f62465b5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.216188] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease: (returnval){ [ 895.216188] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc5d36-bf76-7908-fc7c-ab5d033d989e" [ 895.216188] env[62558]: _type = "HttpNfcLease" [ 895.216188] env[62558]: } obtained for exporting VM: (result){ [ 895.216188] env[62558]: value = "vm-272561" [ 895.216188] env[62558]: _type = "VirtualMachine" [ 895.216188] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 895.216435] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the lease: (returnval){ [ 895.216435] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc5d36-bf76-7908-fc7c-ab5d033d989e" [ 895.216435] env[62558]: _type = "HttpNfcLease" [ 895.216435] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 895.223419] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.223419] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc5d36-bf76-7908-fc7c-ab5d033d989e" [ 895.223419] env[62558]: _type = "HttpNfcLease" [ 895.223419] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 895.312051] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.312440] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc9ab174-2606-461a-9e67-67e939b8c697 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.323950] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266923, 'name': Rename_Task, 'duration_secs': 0.179794} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.325318] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.326596] env[62558]: DEBUG oslo_vmware.api [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 895.326596] env[62558]: value = "task-1266925" [ 895.326596] env[62558]: _type = "Task" [ 895.326596] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.326922] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5896ca8-93d1-486f-b24a-e0d704da1f60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.337916] env[62558]: DEBUG oslo_vmware.api [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.339581] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 895.339581] env[62558]: value = "task-1266926" [ 895.339581] env[62558]: _type = "Task" [ 895.339581] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.348572] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.635114] env[62558]: INFO nova.compute.resource_tracker [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating resource usage from migration d5cc95be-a6ee-4246-bcb7-88ff1cd2f615 [ 895.725288] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.725288] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc5d36-bf76-7908-fc7c-ab5d033d989e" [ 895.725288] env[62558]: _type = "HttpNfcLease" [ 895.725288] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 895.725288] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 895.725288] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc5d36-bf76-7908-fc7c-ab5d033d989e" [ 895.725288] env[62558]: _type = "HttpNfcLease" [ 895.725288] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 895.725949] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cff9f3-a717-4043-80bd-b6dd2d935582 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.735912] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 895.736121] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 895.841018] env[62558]: DEBUG oslo_vmware.api [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266925, 'name': PowerOffVM_Task, 'duration_secs': 0.199717} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.841018] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.841018] env[62558]: DEBUG nova.compute.manager [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.842618] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eae7805-3e38-45cb-a151-f0083b600182 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.854430] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d43dbd7e-d02f-431b-8aa2-e4de33cc6ad9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.858053] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266926, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.008827] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b420f5d7-505e-49d1-87b1-4f45177af403 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.016489] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589bcb62-4fcf-4675-bf97-a6d20732ff21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.054389] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a02398-7f85-46aa-b7f5-29b960d5adc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.062594] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5af1e8-8af2-42c7-b836-31864d90d613 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.077277] env[62558]: DEBUG nova.compute.provider_tree [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.353066] env[62558]: DEBUG oslo_vmware.api [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266926, 'name': PowerOnVM_Task, 'duration_secs': 0.633242} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.353538] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.353838] env[62558]: INFO nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Took 9.04 seconds to spawn the instance on the hypervisor. [ 896.354248] env[62558]: DEBUG nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.355193] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd08cc0-1746-4713-805f-9a4bf3e22765 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.363341] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e117fe25-5d1b-4b9c-80d6-80974d2f32d4 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.572s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.580791] env[62558]: DEBUG nova.scheduler.client.report [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.881938] env[62558]: INFO nova.compute.manager [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Took 34.02 seconds to build instance. [ 897.092155] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.473s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.092155] env[62558]: INFO nova.compute.manager [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Migrating [ 897.092155] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.092155] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.094350] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.016s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.096177] env[62558]: INFO nova.compute.claims [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.100933] env[62558]: INFO nova.compute.rpcapi [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 897.101893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.384780] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a5e046d2-43ee-4b61-9e55-75b24c8d3058 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.537s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.514501] env[62558]: INFO nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Rebuilding instance [ 897.569817] env[62558]: DEBUG nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.571551] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd9fac9-dcd9-4615-8912-95edb7f7a9a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.619780] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.619970] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.620163] env[62558]: DEBUG nova.network.neutron [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.083498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "016c31f6-648c-41f5-909f-5a1ae6366b98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.083817] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.085029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.085029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.085029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.088091] env[62558]: INFO nova.compute.manager [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Terminating instance [ 898.090275] env[62558]: DEBUG nova.compute.manager [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.090461] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.091387] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47daa5d4-059a-4eb4-804c-120b8017caa7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.099293] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.099930] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7b8ad1a-177e-4d12-ab4d-1e4cbb74d999 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.111547] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 898.111547] env[62558]: value = "task-1266928" [ 898.111547] env[62558]: _type = "Task" [ 898.111547] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.111836] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.112190] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18e07cb3-427f-4be6-85f6-386fcacec616 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.127175] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 898.127496] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.127791] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 898.127791] env[62558]: value = "task-1266929" [ 898.127791] env[62558]: _type = "Task" [ 898.127791] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.131799] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a234b143-b135-40af-9394-30592f937a60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.142620] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.144906] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.145182] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e4774be-fcc0-4518-980e-a7ead21cc69b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.211646] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.211794] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.212441] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.215458] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4026fbd9-fda0-4a97-9087-917d0fd7d091 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.225203] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 898.225203] env[62558]: value = "task-1266931" [ 898.225203] env[62558]: _type = "Task" [ 898.225203] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.236195] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.416089] env[62558]: DEBUG nova.network.neutron [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.459017] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f930290-e7bd-4028-b702-d9b5c8e92bdf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.466235] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72943ac1-02b1-4b33-b9d0-dc68ca2affea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.513161] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb838974-3e93-4a64-b8f6-c4ccbf0a4f06 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.523608] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe51c2b-5224-4cba-8df8-d57d27c6ff31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.542096] env[62558]: DEBUG nova.compute.provider_tree [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.645255] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266929, 'name': PowerOffVM_Task, 'duration_secs': 0.222841} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.645255] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.645255] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.645553] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2e16310-062c-4ab2-8f73-c76084a0cbef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.708784] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.709008] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.709199] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore2] 016c31f6-648c-41f5-909f-5a1ae6366b98 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.709474] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae49a6fb-aa6b-4410-94dd-f8cf83435c80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.718151] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 898.718151] env[62558]: value = "task-1266933" [ 898.718151] env[62558]: _type = "Task" [ 898.718151] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.726785] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.735698] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139999} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.735793] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.735994] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.736184] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.923975] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.047578] env[62558]: DEBUG nova.scheduler.client.report [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.228277] env[62558]: DEBUG oslo_vmware.api [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1266933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146387} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.228941] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.229315] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.229641] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.229925] env[62558]: INFO nova.compute.manager [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Took 1.14 seconds to destroy the instance on the hypervisor. [ 899.230314] env[62558]: DEBUG oslo.service.loopingcall [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.232627] env[62558]: DEBUG nova.compute.manager [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.232627] env[62558]: DEBUG nova.network.neutron [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 899.552998] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.553700] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.557382] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.173s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.558138] env[62558]: INFO nova.compute.claims [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.678370] env[62558]: DEBUG nova.compute.manager [req-3ad0d92d-5473-45e1-9a35-68e18cb39c9d req-0eb524d1-f31d-4098-8fb5-167d23e5e2b7 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Received event network-vif-deleted-031eaee8-ff01-4bbf-8d37-5db32153834f {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.678370] env[62558]: INFO nova.compute.manager [req-3ad0d92d-5473-45e1-9a35-68e18cb39c9d req-0eb524d1-f31d-4098-8fb5-167d23e5e2b7 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Neutron deleted interface 031eaee8-ff01-4bbf-8d37-5db32153834f; detaching it from the instance and deleting it from the info cache [ 899.678370] env[62558]: DEBUG nova.network.neutron [req-3ad0d92d-5473-45e1-9a35-68e18cb39c9d req-0eb524d1-f31d-4098-8fb5-167d23e5e2b7 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.775400] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.776113] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.776113] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.776113] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.777664] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.777887] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.778133] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.778308] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.778816] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.778816] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.778816] env[62558]: DEBUG nova.virt.hardware [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.779722] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4dd28f-1cce-462e-80ae-ab2bb5fa42a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.789371] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f642911c-bb9e-4187-8e01-65c45cb6e793" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.789629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.796177] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11702091-66b1-45bb-8658-100aef03476d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.811312] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:bb:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ba08c86-31c9-4715-a527-8ae92d12af2e', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.818770] env[62558]: DEBUG oslo.service.loopingcall [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.819367] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.819618] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e88386b-a660-4b0f-ac28-4fe7d8c953e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.839351] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.839351] env[62558]: value = "task-1266934" [ 899.839351] env[62558]: _type = "Task" [ 899.839351] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.847198] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266934, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.009321] env[62558]: DEBUG nova.network.neutron [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.066427] env[62558]: DEBUG nova.compute.utils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.067766] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.068022] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 900.076148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.076148] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.153798] env[62558]: DEBUG nova.policy [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'daeec08006cc4074ac5dc398309e7d47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0182f4eec034883bfac1e6c85069265', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.179874] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74dec07d-bc0a-495e-8fc4-8a2878520728 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.192462] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56570b5c-6e06-4a6e-a0d2-91c085f2bdff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.224988] env[62558]: DEBUG nova.compute.manager [req-3ad0d92d-5473-45e1-9a35-68e18cb39c9d req-0eb524d1-f31d-4098-8fb5-167d23e5e2b7 service nova] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Detach interface failed, port_id=031eaee8-ff01-4bbf-8d37-5db32153834f, reason: Instance 016c31f6-648c-41f5-909f-5a1ae6366b98 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 900.259233] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f8d6df86-52b8-4692-add4-1ffec82cc598" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.259611] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.293200] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.349992] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266934, 'name': CreateVM_Task, 'duration_secs': 0.479416} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.350204] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.351209] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.351382] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.352183] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.352475] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c509695-3170-4510-87b0-7aef5cf3761f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.357726] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 900.357726] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526680f6-dcfd-904e-28de-b89ced3a5f0b" [ 900.357726] env[62558]: _type = "Task" [ 900.357726] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.368983] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526680f6-dcfd-904e-28de-b89ced3a5f0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.442427] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c0d8cf-a973-4bcb-8b31-b591c3937b9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.464162] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 900.511382] env[62558]: INFO nova.compute.manager [-] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Took 1.28 seconds to deallocate network for instance. [ 900.576424] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.579769] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.641544] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Successfully created port: d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.676189] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.676459] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.766882] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.814885] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.869116] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526680f6-dcfd-904e-28de-b89ced3a5f0b, 'name': SearchDatastore_Task, 'duration_secs': 0.010246} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.869635] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.869752] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.869903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.870057] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.870242] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.870543] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c2746f3-418d-4e3c-be17-31aebf9f3986 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.880359] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.880587] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.881441] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17cfff4e-c722-45d2-9899-3a2c6e8324bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.886930] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 900.886930] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5268a680-4937-2a05-1673-a39b8b9b6c3c" [ 900.886930] env[62558]: _type = "Task" [ 900.886930] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.898951] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5268a680-4937-2a05-1673-a39b8b9b6c3c, 'name': SearchDatastore_Task, 'duration_secs': 0.008051} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.899682] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f7a8fe6-70b7-40d2-acec-37e25da52130 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.906316] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 900.906316] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5226babb-4e44-e0e2-8c76-9bf5b3559062" [ 900.906316] env[62558]: _type = "Task" [ 900.906316] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.914220] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5226babb-4e44-e0e2-8c76-9bf5b3559062, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.916947] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf49701-0cfe-44ae-8b81-164594c45cc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.924212] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca70230c-9a7d-41ba-a14e-025fd6c34d7a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.958717] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b302372a-d6ce-479b-9d00-5f24269284aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.967228] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95071c08-b54c-4323-99dc-0232ba3e1738 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.973283] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.973654] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48f9c62f-3626-4255-b772-dce3e66ac49e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.989636] env[62558]: DEBUG nova.compute.provider_tree [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.993053] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 900.993053] env[62558]: value = "task-1266936" [ 900.993053] env[62558]: _type = "Task" [ 900.993053] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.000553] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.022704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.102188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.179562] env[62558]: INFO nova.compute.manager [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Detaching volume ac9c0aa2-f602-4f72-a5d5-7e7027472d3c [ 901.214739] env[62558]: INFO nova.virt.block_device [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Attempting to driver detach volume ac9c0aa2-f602-4f72-a5d5-7e7027472d3c from mountpoint /dev/sdb [ 901.214900] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 901.215136] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272538', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'name': 'volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1d242e6-1561-4bd4-8e39-281ab6346661', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'serial': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 901.216097] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e162cf-4f77-4e61-aa0f-98c3063636e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.239178] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb910f50-6087-4cc6-8699-3dacf5f36628 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.246768] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e833c416-a3c2-4f91-bcf5-c1aa98e38f08 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.269696] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e012782-a022-4caa-a818-a6dfe1923843 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.291208] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] The volume has not been displaced from its original location: [datastore2] volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c/volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 901.296840] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfiguring VM instance instance-0000003c to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 901.298542] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.299332] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dcb586d-bb44-4037-89d5-e60158cd92a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.318415] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 901.318415] env[62558]: value = "task-1266937" [ 901.318415] env[62558]: _type = "Task" [ 901.318415] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.327494] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.417085] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5226babb-4e44-e0e2-8c76-9bf5b3559062, 'name': SearchDatastore_Task, 'duration_secs': 0.008628} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.417381] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.417702] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.418026] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-daaaae9b-1e9e-4d53-a434-c92c86970474 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.424774] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 901.424774] env[62558]: value = "task-1266938" [ 901.424774] env[62558]: _type = "Task" [ 901.424774] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.435226] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.494694] env[62558]: DEBUG nova.scheduler.client.report [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.507848] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266936, 'name': PowerOffVM_Task, 'duration_secs': 0.249711} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.508168] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.508343] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 901.584690] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.612808] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.613107] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.613270] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.613463] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.613644] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.613799] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.614016] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.614185] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.614424] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.614516] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.614692] env[62558]: DEBUG nova.virt.hardware [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.615640] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a713534-d589-46e6-8cf3-c18579e92746 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.624280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940f9429-9ba5-4ca8-bac8-69bf2971540e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.832636] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266937, 'name': ReconfigVM_Task, 'duration_secs': 0.253175} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.832964] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Reconfigured VM instance instance-0000003c to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 901.837843] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-331c4c2f-131e-4a65-9300-dd181f3b7ec7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.855902] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 901.855902] env[62558]: value = "task-1266939" [ 901.855902] env[62558]: _type = "Task" [ 901.855902] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.864963] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266939, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.934711] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423785} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.935092] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.935605] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.935674] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2878265-45a5-4ae9-94c8-143ee42fdf07 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.943324] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 901.943324] env[62558]: value = "task-1266940" [ 901.943324] env[62558]: _type = "Task" [ 901.943324] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.952431] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.003353] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.003946] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.006521] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.385s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.006714] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.009978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.274s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.009978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.011940] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.156s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.011940] env[62558]: DEBUG nova.objects.instance [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lazy-loading 'resources' on Instance uuid 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.015941] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.016167] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.016419] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.017915] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.026062] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca1bce9a-ebaf-4af3-aa66-cadc1bea5202 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.038431] env[62558]: INFO nova.scheduler.client.report [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted allocations for instance 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b [ 902.048816] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 902.048816] env[62558]: value = "task-1266941" [ 902.048816] env[62558]: _type = "Task" [ 902.048816] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.059514] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266941, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.122120] env[62558]: DEBUG nova.compute.manager [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-vif-plugged-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.122120] env[62558]: DEBUG oslo_concurrency.lockutils [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] Acquiring lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.122120] env[62558]: DEBUG oslo_concurrency.lockutils [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.122120] env[62558]: DEBUG oslo_concurrency.lockutils [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.124876] env[62558]: DEBUG nova.compute.manager [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] No waiting events found dispatching network-vif-plugged-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.124876] env[62558]: WARNING nova.compute.manager [req-797992b5-3f3e-42d1-9e44-492d41e959f6 req-f4f0b9f8-a049-4dde-939e-34d0e327aea2 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received unexpected event network-vif-plugged-d6d6deff-eea7-4653-854b-69c6e43b962a for instance with vm_state building and task_state spawning. [ 902.190854] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Successfully updated port: d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.366733] env[62558]: DEBUG oslo_vmware.api [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266939, 'name': ReconfigVM_Task, 'duration_secs': 0.148692} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.367170] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272538', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'name': 'volume-ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1d242e6-1561-4bd4-8e39-281ab6346661', 'attached_at': '', 'detached_at': '', 'volume_id': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c', 'serial': 'ac9c0aa2-f602-4f72-a5d5-7e7027472d3c'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 902.453364] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063839} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.453642] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.454453] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6dcf120-2fd0-4ef0-a168-a216100e9735 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.478417] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.478577] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8d19f94-11bc-472c-91c1-8fd96c832e14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.500716] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 902.500716] env[62558]: value = "task-1266943" [ 902.500716] env[62558]: _type = "Task" [ 902.500716] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.508828] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.514372] env[62558]: DEBUG nova.compute.utils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.520081] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 902.520081] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.524079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-28918d04-5722-4e41-9a29-4299314ec73b tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.205s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.525586] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 13.733s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.525586] env[62558]: INFO nova.compute.manager [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Unshelving [ 902.558367] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e45f8668-1927-455e-8933-278716a7a576 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "0aeb55df-48e8-4314-ab47-05ff0b5eaa7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.119s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.567270] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266941, 'name': ReconfigVM_Task, 'duration_secs': 0.431409} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.567651] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 902.576599] env[62558]: DEBUG nova.policy [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.694100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.694285] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.694443] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.817582] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8cf655-2dc8-4834-b74f-dfdcfbb32b54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.827973] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f146691b-89a4-4eee-a5fc-91a2a398b6b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.865050] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Successfully created port: f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.867688] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d9fb02-80c6-4579-86c0-7d6ac569b22c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.876883] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c497c5-09f1-4237-95df-47a3555cdc02 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.893903] env[62558]: DEBUG nova.compute.provider_tree [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.918292] env[62558]: DEBUG nova.objects.instance [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.011502] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266943, 'name': ReconfigVM_Task, 'duration_secs': 0.490439} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.011810] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed/6c2fedb1-6240-482d-9ad2-e8dabde4d7ed.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.013153] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-244b9653-8016-4a2b-a49c-9aab97080713 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.018520] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.022851] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 903.022851] env[62558]: value = "task-1266944" [ 903.022851] env[62558]: _type = "Task" [ 903.022851] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.036641] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266944, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.079716] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.079716] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.079994] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.080056] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.080214] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.080363] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.080582] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.080782] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.080963] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.081148] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.081327] env[62558]: DEBUG nova.virt.hardware [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.086839] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 903.089253] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a925b40-6a2e-4fce-9bd4-8322b7dfa149 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.112671] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 903.112671] env[62558]: value = "task-1266945" [ 903.112671] env[62558]: _type = "Task" [ 903.112671] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.122040] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266945, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.253193] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.398547] env[62558]: DEBUG nova.scheduler.client.report [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.473892] env[62558]: DEBUG nova.network.neutron [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.541175] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266944, 'name': Rename_Task, 'duration_secs': 0.19646} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.541175] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.541175] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8f4ada2-df26-4d2f-8d8d-0d0d199c0098 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.549506] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 903.549506] env[62558]: value = "task-1266946" [ 903.549506] env[62558]: _type = "Task" [ 903.549506] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.560828] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.562828] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.623238] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266945, 'name': ReconfigVM_Task, 'duration_secs': 0.219206} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.623238] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 903.623849] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b18f26-9ce2-4e06-bb44-f82851fd95f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.647358] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.647738] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89978420-8098-4eca-9768-af0292af89b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.667574] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 903.667574] env[62558]: value = "task-1266947" [ 903.667574] env[62558]: _type = "Task" [ 903.667574] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.676717] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266947, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.903681] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.904566] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.907214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.042s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.907506] env[62558]: DEBUG nova.objects.instance [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'resources' on Instance uuid 4526df2a-f801-4de8-8218-497a7d22034f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.926454] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4f1a9f97-5980-4c26-ae9c-49bcdb7973e6 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.250s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.927568] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.024s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.928273] env[62558]: DEBUG nova.compute.manager [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.929068] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a32d645-dd90-4fc7-a8fc-f52c6390c886 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.937959] env[62558]: DEBUG nova.compute.manager [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 903.938635] env[62558]: DEBUG nova.objects.instance [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.946646] env[62558]: INFO nova.scheduler.client.report [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted allocations for instance 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca [ 903.979031] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.979031] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Instance network_info: |[{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.979031] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:3f:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '535b175f-71d3-4226-81ab-ca253f27fedd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6d6deff-eea7-4653-854b-69c6e43b962a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.984980] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Creating folder: Project (a0182f4eec034883bfac1e6c85069265). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 903.985314] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a11d432-ef7c-4599-918d-a2d13b3e042b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.002830] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Created folder: Project (a0182f4eec034883bfac1e6c85069265) in parent group-v272451. [ 904.002830] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Creating folder: Instances. Parent ref: group-v272564. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.003257] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6fd4bbf-3965-4aa9-aa98-ad0914133389 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.016030] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 904.017529] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039a02e7-1bf8-4863-9527-9fffa8f174cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.022330] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Created folder: Instances in parent group-v272564. [ 904.023068] env[62558]: DEBUG oslo.service.loopingcall [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.023233] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.023500] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35e025c9-b53e-422d-b27a-0fd6e33c1124 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.041654] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.044035] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 904.044226] env[62558]: ERROR oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk due to incomplete transfer. [ 904.045159] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4b5608a8-02ad-413a-91ac-bcffe9d7b9ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.051169] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.051169] env[62558]: value = "task-1266950" [ 904.051169] env[62558]: _type = "Task" [ 904.051169] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.058591] env[62558]: DEBUG oslo_vmware.rw_handles [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52874345-1aa6-a1bf-2895-f0b1ee48d938/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 904.058803] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Uploaded image fe8845c5-445d-4f71-8ce6-a1a0ce770a67 to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 904.060992] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 904.063676] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fec1d903-8eef-4f28-a207-fdd6b53fcf2c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.069461] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.072043] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.072043] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.072250] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.072383] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.072579] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.072741] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.072963] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.073168] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.073352] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.073536] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.073718] env[62558]: DEBUG nova.virt.hardware [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.079197] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82d9d96-9a8e-4d05-95a5-de1553eedeb6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.081970] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 904.081970] env[62558]: value = "task-1266951" [ 904.081970] env[62558]: _type = "Task" [ 904.081970] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.082256] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266950, 'name': CreateVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.089032] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56df6294-1732-4288-b5d8-7dfff8d8b9bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.098114] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266951, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.148419] env[62558]: DEBUG nova.compute.manager [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.148662] env[62558]: DEBUG nova.compute.manager [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing instance network info cache due to event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.148915] env[62558]: DEBUG oslo_concurrency.lockutils [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.149493] env[62558]: DEBUG oslo_concurrency.lockutils [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.149591] env[62558]: DEBUG nova.network.neutron [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.178702] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.444179] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.444538] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ba48d90-9322-47ad-94c0-b45527e7aae6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.457998] env[62558]: DEBUG oslo_vmware.api [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 904.457998] env[62558]: value = "task-1266953" [ 904.457998] env[62558]: _type = "Task" [ 904.457998] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.458513] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0542f998-5884-4477-b616-c7c2ec935674 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "2a7b4e08-d9b1-49f1-9611-ca1acc3315ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.649s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.469923] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Successfully updated port: f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.475996] env[62558]: DEBUG oslo_vmware.api [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.567431] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266946, 'name': PowerOnVM_Task, 'duration_secs': 0.80242} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.567754] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.568033] env[62558]: DEBUG nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.568885] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee4ddd4-efe2-40df-b643-17c26ed1f317 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.575073] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266950, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.598960] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266951, 'name': Destroy_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.679807] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266947, 'name': ReconfigVM_Task, 'duration_secs': 0.570918} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.679807] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7/63050323-047e-4d73-91ae-859467b4b5a7.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.679988] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 904.711223] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2216971-4d95-447b-9ed7-1db616064b0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.719870] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ad01a9-1643-432f-92f3-1672f474142a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.758577] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89951eee-8346-47b0-8981-8ba002bd3308 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.768052] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8a2bcc-e0f7-4cc5-9010-f9930635b78d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.781859] env[62558]: DEBUG nova.compute.provider_tree [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.918746] env[62558]: DEBUG nova.network.neutron [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updated VIF entry in instance network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.919180] env[62558]: DEBUG nova.network.neutron [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.972051] env[62558]: DEBUG oslo_vmware.api [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266953, 'name': PowerOffVM_Task, 'duration_secs': 0.511551} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.972371] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.972599] env[62558]: DEBUG nova.compute.manager [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.973489] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526e75db-0f04-4fb0-b65a-ad1bb1619379 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.980432] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.980590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.981169] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.067983] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266950, 'name': CreateVM_Task, 'duration_secs': 0.671095} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.068299] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.069290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.069458] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.069783] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.070065] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46195632-dfb3-4bb7-a375-17b953ce5f7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.075980] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 905.075980] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527967fd-6e87-3d2c-8249-39e3cc916b08" [ 905.075980] env[62558]: _type = "Task" [ 905.075980] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.084703] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527967fd-6e87-3d2c-8249-39e3cc916b08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.094442] env[62558]: INFO nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] bringing vm to original state: 'stopped' [ 905.096877] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266951, 'name': Destroy_Task, 'duration_secs': 0.903681} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.097293] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Destroyed the VM [ 905.097752] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 905.098014] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5ddfc2d0-af26-454d-90d7-5146833f2b16 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.105899] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 905.105899] env[62558]: value = "task-1266954" [ 905.105899] env[62558]: _type = "Task" [ 905.105899] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.114632] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266954, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.191074] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd5ddee-a9fb-40c4-b60d-ab2867379953 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.216341] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c43fc7-8f9f-450d-8563-1899744e5e6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.238022] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 905.287045] env[62558]: DEBUG nova.scheduler.client.report [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.426234] env[62558]: DEBUG oslo_concurrency.lockutils [req-6c775eb5-e564-4536-96a0-8409b3c9d6eb req-c0b59212-e66b-4fdc-8481-adea9b87324c service nova] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.488789] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a2f8b43a-54cc-400e-9dd9-e5420da09229 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.561s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.516734] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.591721] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527967fd-6e87-3d2c-8249-39e3cc916b08, 'name': SearchDatastore_Task, 'duration_secs': 0.029909} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.592082] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.592341] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.592719] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.592907] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.593117] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.593399] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c9db19c-d7d3-48c6-b061-f8d1bf018ab8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.612831] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.613204] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.614220] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61351f55-4174-4a9a-828d-98d5495fa19d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.621118] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266954, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.625166] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 905.625166] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ad147-353d-1e11-2607-e2e42f9a18b8" [ 905.625166] env[62558]: _type = "Task" [ 905.625166] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.634788] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ad147-353d-1e11-2607-e2e42f9a18b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.689028] env[62558]: DEBUG nova.network.neutron [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Updating instance_info_cache with network_info: [{"id": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "address": "fa:16:3e:41:1a:98", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d0d5f5-cb", "ovs_interfaceid": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.778224] env[62558]: DEBUG nova.network.neutron [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Port 174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 905.791670] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.795150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.275s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.795150] env[62558]: DEBUG nova.objects.instance [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lazy-loading 'resources' on Instance uuid a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.813171] env[62558]: INFO nova.scheduler.client.report [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted allocations for instance 4526df2a-f801-4de8-8218-497a7d22034f [ 906.104771] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.105115] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.105311] env[62558]: DEBUG nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.106215] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12854d2b-fa25-46b5-b71e-dbe813a871fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.119062] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266954, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.120870] env[62558]: DEBUG nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 906.123183] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.123401] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a3e6ffa-3787-45b1-baa2-91af79208db6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.131622] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 906.131622] env[62558]: value = "task-1266955" [ 906.131622] env[62558]: _type = "Task" [ 906.131622] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.135200] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ad147-353d-1e11-2607-e2e42f9a18b8, 'name': SearchDatastore_Task, 'duration_secs': 0.011197} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.139169] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24c02ccc-4a5d-4233-a116-ebf2d7013419 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.147478] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.148956] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 906.148956] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5265e450-8107-5181-65e4-855a5c134bc9" [ 906.148956] env[62558]: _type = "Task" [ 906.148956] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.159008] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5265e450-8107-5181-65e4-855a5c134bc9, 'name': SearchDatastore_Task, 'duration_secs': 0.010708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.159311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.159617] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/7aebdb8a-517b-4168-91e0-1b704e6a11c7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 906.159927] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04dd1645-9876-441a-aa06-495515b0fcd0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.168230] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 906.168230] env[62558]: value = "task-1266956" [ 906.168230] env[62558]: _type = "Task" [ 906.168230] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.177419] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.180070] env[62558]: DEBUG nova.compute.manager [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Received event network-vif-plugged-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.180379] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Acquiring lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.180686] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.180817] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.180987] env[62558]: DEBUG nova.compute.manager [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] No waiting events found dispatching network-vif-plugged-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.181173] env[62558]: WARNING nova.compute.manager [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Received unexpected event network-vif-plugged-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 for instance with vm_state building and task_state spawning. [ 906.181336] env[62558]: DEBUG nova.compute.manager [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Received event network-changed-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.181492] env[62558]: DEBUG nova.compute.manager [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Refreshing instance network info cache due to event network-changed-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.181662] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Acquiring lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.193205] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.193638] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Instance network_info: |[{"id": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "address": "fa:16:3e:41:1a:98", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d0d5f5-cb", "ovs_interfaceid": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 906.194154] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Acquired lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.194241] env[62558]: DEBUG nova.network.neutron [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Refreshing network info cache for port f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.195521] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:1a:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f78b07ea-f425-4622-84f4-706a5d8820a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0d0d5f5-cb9e-499e-ad0c-3025f93b4655', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.203017] env[62558]: DEBUG oslo.service.loopingcall [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.205613] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.207724] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e59cf9a-34d1-4a89-8d8e-58981a09eeb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.225176] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.225473] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.235404] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.235404] env[62558]: value = "task-1266957" [ 906.235404] env[62558]: _type = "Task" [ 906.235404] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.247695] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266957, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.322515] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b7813fe4-68b3-4510-8185-574b3b16494b tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "4526df2a-f801-4de8-8218-497a7d22034f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.913s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.431708] env[62558]: DEBUG nova.objects.instance [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.603717] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50524751-dcd8-4447-a039-91ab8e61e13c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.617242] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e6d5a5-e2dc-4f47-9615-9b3ee59ecbba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.627068] env[62558]: DEBUG oslo_vmware.api [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1266954, 'name': RemoveSnapshot_Task, 'duration_secs': 1.162115} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.672388] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 906.672804] env[62558]: INFO nova.compute.manager [None req-16a2c207-d2f6-466e-90c5-9ccf67db276b tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Took 16.64 seconds to snapshot the instance on the hypervisor. [ 906.688143] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cb429a-2ff6-443c-b94f-533231c4cee3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.697496] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266956, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471303} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.702424] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/7aebdb8a-517b-4168-91e0-1b704e6a11c7.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.702774] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.703070] env[62558]: DEBUG oslo_vmware.api [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266955, 'name': PowerOffVM_Task, 'duration_secs': 0.204772} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.703300] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef09a840-c6d7-4863-a898-1452e3040507 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.706732] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36efcb23-1f3f-41b4-a72a-4f9fac9f82fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.710438] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.710679] env[62558]: DEBUG nova.compute.manager [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.714102] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0adcd3a-6133-4890-b67c-67d549ac5612 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.733612] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.736663] env[62558]: DEBUG nova.compute.provider_tree [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.737806] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 906.737806] env[62558]: value = "task-1266958" [ 906.737806] env[62558]: _type = "Task" [ 906.737806] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.757467] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.757753] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266957, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.805517] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.805759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.805943] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.933206] env[62558]: DEBUG nova.network.neutron [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Updated VIF entry in instance network info cache for port f0d0d5f5-cb9e-499e-ad0c-3025f93b4655. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.933771] env[62558]: DEBUG nova.network.neutron [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Updating instance_info_cache with network_info: [{"id": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "address": "fa:16:3e:41:1a:98", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d0d5f5-cb", "ovs_interfaceid": "f0d0d5f5-cb9e-499e-ad0c-3025f93b4655", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.939662] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.939662] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.939662] env[62558]: DEBUG nova.network.neutron [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.939662] env[62558]: DEBUG nova.objects.instance [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'info_cache' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.241038] env[62558]: DEBUG nova.scheduler.client.report [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.249123] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.144s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.264934] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078005} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.268186] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.268487] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266957, 'name': CreateVM_Task, 'duration_secs': 0.541512} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.269625] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0bd9c8-af70-45ee-a0b7-49a3673eb973 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.272094] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.273061] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.274094] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.274263] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.274575] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.275225] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5aac4e63-bd83-45d6-9b57-3af3a9963726 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.280576] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 907.280576] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a17b3e-f4b5-43ce-a87a-352ffeced666" [ 907.280576] env[62558]: _type = "Task" [ 907.280576] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.298297] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/7aebdb8a-517b-4168-91e0-1b704e6a11c7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.301598] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4f61044-d165-4f1a-baa7-04b44e748337 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.323750] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a17b3e-f4b5-43ce-a87a-352ffeced666, 'name': SearchDatastore_Task, 'duration_secs': 0.010361} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.326344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.326344] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.326344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.326344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.326514] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.326776] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 907.326776] env[62558]: value = "task-1266959" [ 907.326776] env[62558]: _type = "Task" [ 907.326776] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.326957] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a2c575f-e99f-4ed8-814b-8f077eb4e3fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.339245] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266959, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.340684] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.340865] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.341579] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d3c1c64-4c07-4dc6-af38-d62ae8f51d8a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.348192] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 907.348192] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5270d363-45bc-443c-c3f3-31a2564c403c" [ 907.348192] env[62558]: _type = "Task" [ 907.348192] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.355907] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5270d363-45bc-443c-c3f3-31a2564c403c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.438677] env[62558]: DEBUG oslo_concurrency.lockutils [req-a73b1b74-f209-422e-83c6-5aafdaddf557 req-5fe4f481-b04c-4acc-8856-98fe9c1f6f44 service nova] Releasing lock "refresh_cache-ecf01ab6-5019-4453-b102-0e754abc4ef8" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.441621] env[62558]: DEBUG nova.objects.base [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 907.565684] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.565939] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.750378] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.956s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.753190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.938s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.755008] env[62558]: INFO nova.compute.claims [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.767850] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.781706] env[62558]: INFO nova.scheduler.client.report [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Deleted allocations for instance a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884 [ 907.845027] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266959, 'name': ReconfigVM_Task, 'duration_secs': 0.322305} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.845027] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/7aebdb8a-517b-4168-91e0-1b704e6a11c7.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.845027] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6054fd61-a59e-4eb5-b1aa-e76ab8b47c1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.853525] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 907.853525] env[62558]: value = "task-1266960" [ 907.853525] env[62558]: _type = "Task" [ 907.853525] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.860842] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5270d363-45bc-443c-c3f3-31a2564c403c, 'name': SearchDatastore_Task, 'duration_secs': 0.014263} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.861934] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-120aa5dc-daa2-4ce9-a681-7961b3da718a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.867685] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266960, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.871123] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 907.871123] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529c6db9-54c2-50a2-efd6-46acc827da42" [ 907.871123] env[62558]: _type = "Task" [ 907.871123] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.875758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.876038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.876337] env[62558]: DEBUG nova.network.neutron [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.880474] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529c6db9-54c2-50a2-efd6-46acc827da42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.068361] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.192043] env[62558]: DEBUG nova.network.neutron [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [{"id": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "address": "fa:16:3e:a3:00:90", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdbc998f-ca", "ovs_interfaceid": "bdbc998f-cabe-438b-b3ce-6477b78a13e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.288559] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f149ab86-8831-408f-ae9d-c8ccd7b84405 tempest-ListServersNegativeTestJSON-76059707 tempest-ListServersNegativeTestJSON-76059707-project-member] Lock "a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.339s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.364768] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266960, 'name': Rename_Task, 'duration_secs': 0.156245} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.365715] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.365997] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fae6c97c-77b0-481b-af4f-499b3c1be4cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.376827] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 908.376827] env[62558]: value = "task-1266961" [ 908.376827] env[62558]: _type = "Task" [ 908.376827] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.386426] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529c6db9-54c2-50a2-efd6-46acc827da42, 'name': SearchDatastore_Task, 'duration_secs': 0.013703} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.387368] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.387663] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ecf01ab6-5019-4453-b102-0e754abc4ef8/ecf01ab6-5019-4453-b102-0e754abc4ef8.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.388201] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9587aa4-9911-4e20-95ab-76be28507c4e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.395033] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266961, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.396435] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 908.396435] env[62558]: value = "task-1266962" [ 908.396435] env[62558]: _type = "Task" [ 908.396435] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.407312] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266962, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.596708] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.601741] env[62558]: DEBUG nova.network.neutron [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.696093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "refresh_cache-a1d242e6-1561-4bd4-8e39-281ab6346661" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.893190] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266961, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.908243] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266962, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.061504] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f29ae5b-0d1d-408f-9c23-562df9a08a37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.070777] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb21d28-d168-4a5c-8762-150c97d57626 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.106025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.112019] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196b0e98-7ae4-416a-afe9-84f72bca71a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.118604] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e6f0fe-8006-46e3-800d-dbb5f9fddc6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.134908] env[62558]: DEBUG nova.compute.provider_tree [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.198834] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.199115] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-250b1dcc-0abf-4ade-86e9-40fdca9fd831 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.209501] env[62558]: DEBUG oslo_vmware.api [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 909.209501] env[62558]: value = "task-1266963" [ 909.209501] env[62558]: _type = "Task" [ 909.209501] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.218593] env[62558]: DEBUG oslo_vmware.api [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.390649] env[62558]: DEBUG oslo_vmware.api [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266961, 'name': PowerOnVM_Task, 'duration_secs': 0.585314} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.390938] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.391697] env[62558]: INFO nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Took 7.81 seconds to spawn the instance on the hypervisor. [ 909.391988] env[62558]: DEBUG nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.392843] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22997b79-ad12-40be-9b39-fd29589ed349 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.416341] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266962, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556557} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.416626] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ecf01ab6-5019-4453-b102-0e754abc4ef8/ecf01ab6-5019-4453-b102-0e754abc4ef8.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.416848] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.417135] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb860b9c-32b0-4850-a065-09f8d2178b8a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.425995] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 909.425995] env[62558]: value = "task-1266964" [ 909.425995] env[62558]: _type = "Task" [ 909.425995] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.436361] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.629778] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc220b5-3da6-43e8-b179-abadfeaee9f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.650569] env[62558]: DEBUG nova.scheduler.client.report [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.655320] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc4c1a6-67da-4ce4-8454-99432448d55c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.668720] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 909.724472] env[62558]: DEBUG oslo_vmware.api [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1266963, 'name': PowerOnVM_Task, 'duration_secs': 0.412893} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.725702] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.725967] env[62558]: DEBUG nova.compute.manager [None req-3eac67cb-2dee-4e92-93e8-1f3fd57ab0e9 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.726998] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54d79c3-7f83-424a-a052-28a04aa87748 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.737331] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.737716] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.738037] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.738393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.738706] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.744234] env[62558]: INFO nova.compute.manager [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Terminating instance [ 909.746580] env[62558]: DEBUG nova.compute.manager [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.747048] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.747815] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dad65f5-2c13-4eac-a3be-0fd5e0d2c981 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.757300] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.757802] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c8e9529-d436-455f-a75f-10a0bec3a77a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.844293] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.844564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.845976] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.846083] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.846331] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore1] 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.846905] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80c1f876-b4a5-40e4-a911-64713c328c7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.855300] env[62558]: DEBUG oslo_vmware.api [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 909.855300] env[62558]: value = "task-1266966" [ 909.855300] env[62558]: _type = "Task" [ 909.855300] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.867624] env[62558]: DEBUG oslo_vmware.api [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.920833] env[62558]: INFO nova.compute.manager [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Took 28.86 seconds to build instance. [ 909.939969] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083021} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.941616] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.943054] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a67f7d-faab-4e6a-a218-8a29513e8a22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.973591] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] ecf01ab6-5019-4453-b102-0e754abc4ef8/ecf01ab6-5019-4453-b102-0e754abc4ef8.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.974623] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3480a0d8-7099-42ca-8eee-054a5e0384e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.001012] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 910.001012] env[62558]: value = "task-1266967" [ 910.001012] env[62558]: _type = "Task" [ 910.001012] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.013503] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266967, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.158704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.158704] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.159522] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.137s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.159750] env[62558]: DEBUG nova.objects.instance [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'resources' on Instance uuid 016c31f6-648c-41f5-909f-5a1ae6366b98 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.176475] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.177038] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c75dd2cf-0426-422f-8bde-7c637bb864f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.187166] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 910.187166] env[62558]: value = "task-1266968" [ 910.187166] env[62558]: _type = "Task" [ 910.187166] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.195878] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.348086] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.368091] env[62558]: DEBUG oslo_vmware.api [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.423960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-05ff80d5-5615-4171-a96c-b0c3b0c9f4ac tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.373s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.517979] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.662927] env[62558]: DEBUG nova.compute.utils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.667891] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 910.710343] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266968, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.872873] env[62558]: DEBUG oslo_vmware.api [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1266966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.539784} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.874030] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.874319] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.874510] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.874687] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.874866] env[62558]: INFO nova.compute.manager [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Took 1.13 seconds to destroy the instance on the hypervisor. [ 910.875124] env[62558]: DEBUG oslo.service.loopingcall [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.875330] env[62558]: DEBUG nova.compute.manager [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.875424] env[62558]: DEBUG nova.network.neutron [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.994295] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c789e4-6a43-478c-ad27-ff86eaeb319a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.002588] env[62558]: INFO nova.compute.manager [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Rescuing [ 911.002837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.002996] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.003239] env[62558]: DEBUG nova.network.neutron [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.008794] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53776417-5835-45f6-a454-9b1b5cf423f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.017941] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266967, 'name': ReconfigVM_Task, 'duration_secs': 0.76965} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.042679] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Reconfigured VM instance instance-00000054 to attach disk [datastore2] ecf01ab6-5019-4453-b102-0e754abc4ef8/ecf01ab6-5019-4453-b102-0e754abc4ef8.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.044050] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c0de462-9da8-4588-8fe8-1d008dd14f79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.046962] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1765c68-094d-457a-b678-99e427d2f11d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.059051] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46189cbf-425e-4bec-8d69-26fa9809efe6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.063511] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 911.063511] env[62558]: value = "task-1266969" [ 911.063511] env[62558]: _type = "Task" [ 911.063511] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.077921] env[62558]: DEBUG nova.compute.provider_tree [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.086170] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266969, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.167314] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.196825] env[62558]: DEBUG oslo_vmware.api [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1266968, 'name': PowerOnVM_Task, 'duration_secs': 0.707593} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.196975] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.197213] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abc344ce-9021-4b2b-9290-640f982bd29b tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance '63050323-047e-4d73-91ae-859467b4b5a7' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 911.278235] env[62558]: DEBUG nova.compute.manager [req-532cb8d4-d186-49e7-90f1-c0b6b7d70537 req-5a4cf013-c948-4358-960e-5c72bd4f4d51 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Received event network-vif-deleted-9ba08c86-31c9-4715-a527-8ae92d12af2e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.278456] env[62558]: INFO nova.compute.manager [req-532cb8d4-d186-49e7-90f1-c0b6b7d70537 req-5a4cf013-c948-4358-960e-5c72bd4f4d51 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Neutron deleted interface 9ba08c86-31c9-4715-a527-8ae92d12af2e; detaching it from the instance and deleting it from the info cache [ 911.278673] env[62558]: DEBUG nova.network.neutron [req-532cb8d4-d186-49e7-90f1-c0b6b7d70537 req-5a4cf013-c948-4358-960e-5c72bd4f4d51 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.578325] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266969, 'name': Rename_Task, 'duration_secs': 0.339683} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.579029] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.579228] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c41acc33-b342-44bb-b363-78c8fc364a9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.582610] env[62558]: DEBUG nova.scheduler.client.report [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.598471] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 911.598471] env[62558]: value = "task-1266970" [ 911.598471] env[62558]: _type = "Task" [ 911.598471] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.613246] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.662076] env[62558]: DEBUG nova.network.neutron [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.754901] env[62558]: DEBUG nova.network.neutron [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.784020] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8990807-8c4d-4a18-a0a7-95bf647a138e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.796879] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a243bc5c-3c8b-442e-bd56-57c2549ba7c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.833296] env[62558]: DEBUG nova.compute.manager [req-532cb8d4-d186-49e7-90f1-c0b6b7d70537 req-5a4cf013-c948-4358-960e-5c72bd4f4d51 service nova] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Detach interface failed, port_id=9ba08c86-31c9-4715-a527-8ae92d12af2e, reason: Instance 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 912.096025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.096025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.994s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.099742] env[62558]: INFO nova.compute.claims [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.128233] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266970, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.136267] env[62558]: INFO nova.scheduler.client.report [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance 016c31f6-648c-41f5-909f-5a1ae6366b98 [ 912.163520] env[62558]: INFO nova.compute.manager [-] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Took 1.29 seconds to deallocate network for instance. [ 912.176994] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.212618] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.213226] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.213549] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.214285] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.214285] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.214509] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.214864] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.216476] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.216476] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.216476] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.216476] env[62558]: DEBUG nova.virt.hardware [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.217723] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad2add0-c455-430e-916b-ed029b057577 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.229447] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6356346-2aef-49fe-bb9d-dc7fcfd06223 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.245277] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.251127] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Creating folder: Project (f7ae843eb19b4eb08ff1144b411f439a). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.251552] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f99ee34e-1e24-41a8-9375-cb585a415fd7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.256418] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.265072] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Created folder: Project (f7ae843eb19b4eb08ff1144b411f439a) in parent group-v272451. [ 912.265311] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Creating folder: Instances. Parent ref: group-v272568. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.269215] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b58c6646-6a14-4a2d-a2ea-ef0419c38435 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.284631] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Created folder: Instances in parent group-v272568. [ 912.289020] env[62558]: DEBUG oslo.service.loopingcall [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.289020] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.289020] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de7ee45a-6cb4-4f91-b00c-ee8567ca77ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.316911] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.316911] env[62558]: value = "task-1266973" [ 912.316911] env[62558]: _type = "Task" [ 912.316911] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.328882] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266973, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.616393] env[62558]: DEBUG oslo_vmware.api [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266970, 'name': PowerOnVM_Task, 'duration_secs': 0.559972} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.616943] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.617162] env[62558]: INFO nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Took 8.58 seconds to spawn the instance on the hypervisor. [ 912.617565] env[62558]: DEBUG nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.618521] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02bc3074-8b46-42b6-b03d-0a049f2d802a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.645248] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5466a67d-e7db-4b7b-8b3c-6e5993706d9b tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "016c31f6-648c-41f5-909f-5a1ae6366b98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.561s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.671547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.814196] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.814762] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ee31594-3cf1-4bb6-ad4a-663f11706779 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.828106] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266973, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.830082] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 912.830082] env[62558]: value = "task-1266974" [ 912.830082] env[62558]: _type = "Task" [ 912.830082] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.841224] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.139336] env[62558]: INFO nova.compute.manager [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Took 27.78 seconds to build instance. [ 913.190413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.191046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.191286] env[62558]: DEBUG nova.compute.manager [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Going to confirm migration 1 {{(pid=62558) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 913.330191] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266973, 'name': CreateVM_Task, 'duration_secs': 0.579024} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.330834] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.333808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.333976] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.334353] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.334719] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b986ceb0-55da-4d42-97c3-84ccab550522 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.345978] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 913.345978] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f3c708-ce8a-4a7a-682c-e9f01368202c" [ 913.345978] env[62558]: _type = "Task" [ 913.345978] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.345978] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266974, 'name': PowerOffVM_Task, 'duration_secs': 0.275405} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.345978] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.349180] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e536fc-e10e-44cf-9bda-be3dfb3aacb6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.374953] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f3c708-ce8a-4a7a-682c-e9f01368202c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.375786] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c911618-1956-404c-b787-374238facf48 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.429785] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.430253] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f303784-245b-44fb-b8c8-dab355316da1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.441597] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 913.441597] env[62558]: value = "task-1266975" [ 913.441597] env[62558]: _type = "Task" [ 913.441597] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.453711] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.552477] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55742813-ec6f-406e-ba78-b0ca0377e1e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.560906] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1123a2f-58bb-46e0-928e-46a7ad690429 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.599541] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f946bf28-c8bb-4cb3-b0e4-2cac8e6fade4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.608395] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451ebca8-5b01-4bbe-91e5-0c910299e046 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.623346] env[62558]: DEBUG nova.compute.provider_tree [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.641243] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f9de00b-a86b-4062-9b96-be7d0dd67b82 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.289s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.822674] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.823443] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.823443] env[62558]: DEBUG nova.network.neutron [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.823443] env[62558]: DEBUG nova.objects.instance [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'info_cache' on Instance uuid 63050323-047e-4d73-91ae-859467b4b5a7 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.860123] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f3c708-ce8a-4a7a-682c-e9f01368202c, 'name': SearchDatastore_Task, 'duration_secs': 0.037041} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.860533] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.860775] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.861177] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.861177] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.861358] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.861658] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50d997ff-3045-457e-bff6-f9fbd6826f6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.875636] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.875850] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.876523] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8fc223f-00ad-4f32-9a4d-9a595e764eb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.883237] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 913.883237] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529155d7-1bb9-7013-a953-c67f43bbec96" [ 913.883237] env[62558]: _type = "Task" [ 913.883237] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.893625] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529155d7-1bb9-7013-a953-c67f43bbec96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.910245] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.910484] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.954138] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 913.954441] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.954714] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.954878] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.955073] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.955360] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01215ea6-c646-463b-943b-aa021ca7c294 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.968274] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.968791] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.969270] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2113caf5-299f-470e-9952-0405fedb79f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.977733] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 913.977733] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52df2b51-851b-4f38-40c1-c95483cacf5d" [ 913.977733] env[62558]: _type = "Task" [ 913.977733] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.987814] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52df2b51-851b-4f38-40c1-c95483cacf5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.126670] env[62558]: DEBUG nova.scheduler.client.report [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.134651] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.135408] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.135408] env[62558]: DEBUG nova.compute.manager [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.136308] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27952f6b-c1a8-42a4-8206-42c0ae0eb5a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.144765] env[62558]: DEBUG nova.compute.manager [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 914.146237] env[62558]: DEBUG nova.objects.instance [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'flavor' on Instance uuid ecf01ab6-5019-4453-b102-0e754abc4ef8 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.397724] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529155d7-1bb9-7013-a953-c67f43bbec96, 'name': SearchDatastore_Task, 'duration_secs': 0.035087} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.397724] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7048e6-4397-457b-abd4-b8d537c2cd14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.403894] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 914.403894] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1ae88-a01b-3108-5bd7-3a04800be958" [ 914.403894] env[62558]: _type = "Task" [ 914.403894] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.412551] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 914.415061] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1ae88-a01b-3108-5bd7-3a04800be958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.494917] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52df2b51-851b-4f38-40c1-c95483cacf5d, 'name': SearchDatastore_Task, 'duration_secs': 0.012506} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.494917] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d15fcbc9-f25c-404f-ac5f-fb991ae42be5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.499095] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 914.499095] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529adf70-cf55-e85c-f55b-718c7c50680b" [ 914.499095] env[62558]: _type = "Task" [ 914.499095] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.509333] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529adf70-cf55-e85c-f55b-718c7c50680b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.632872] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.633343] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.636365] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.338s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.637854] env[62558]: INFO nova.compute.claims [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.651334] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.651642] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a295df76-0ad0-4123-bb14-f4709fc51a41 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.662038] env[62558]: DEBUG oslo_vmware.api [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 914.662038] env[62558]: value = "task-1266976" [ 914.662038] env[62558]: _type = "Task" [ 914.662038] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.675847] env[62558]: DEBUG oslo_vmware.api [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.923443] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1ae88-a01b-3108-5bd7-3a04800be958, 'name': SearchDatastore_Task, 'duration_secs': 0.010056} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.925528] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.925739] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f642911c-bb9e-4187-8e01-65c45cb6e793/f642911c-bb9e-4187-8e01-65c45cb6e793.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.927894] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9ddd9f6-f10d-439f-82b2-a38fdc4a465c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.938207] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 914.938207] env[62558]: value = "task-1266977" [ 914.938207] env[62558]: _type = "Task" [ 914.938207] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.952058] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.955293] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.011451] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529adf70-cf55-e85c-f55b-718c7c50680b, 'name': SearchDatastore_Task, 'duration_secs': 0.010722} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.011733] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.012635] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. {{(pid=62558) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 915.012635] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8b4eec9-c05e-48bf-a53a-4d70b769ed65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.021237] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 915.021237] env[62558]: value = "task-1266978" [ 915.021237] env[62558]: _type = "Task" [ 915.021237] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.031625] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.077922] env[62558]: DEBUG nova.network.neutron [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [{"id": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "address": "fa:16:3e:17:76:fa", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174e00d0-f0", "ovs_interfaceid": "174e00d0-f0e5-4b0a-9ca2-4db08bb7f674", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.144260] env[62558]: DEBUG nova.compute.utils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.152030] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.152030] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.176236] env[62558]: DEBUG oslo_vmware.api [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266976, 'name': PowerOffVM_Task, 'duration_secs': 0.16105} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.176870] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.176870] env[62558]: DEBUG nova.compute.manager [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.177558] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4509906e-a70e-4b0f-97ae-1f49c177d8f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.200312] env[62558]: DEBUG nova.policy [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37bb028d611e42d98319c4afba5b1b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46aa5789fccf4925aa49f9c4050c8463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.359026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.359026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.450895] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266977, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.532477] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266978, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.580862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-63050323-047e-4d73-91ae-859467b4b5a7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.581292] env[62558]: DEBUG nova.objects.instance [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'migration_context' on Instance uuid 63050323-047e-4d73-91ae-859467b4b5a7 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.588999] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Successfully created port: 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.656342] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.700535] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5f444b13-3263-44d1-94af-3eb76b833690 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.565s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.864720] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.949598] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542107} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.949598] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f642911c-bb9e-4187-8e01-65c45cb6e793/f642911c-bb9e-4187-8e01-65c45cb6e793.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.949598] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.949853] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c285ab6-65cb-4e55-b506-348c6f56634b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.964748] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 915.964748] env[62558]: value = "task-1266979" [ 915.964748] env[62558]: _type = "Task" [ 915.964748] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.981350] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266979, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.004949] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66dc9de9-1149-4021-aa68-399b58acaece {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.013142] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb908d62-7fa4-41db-a836-c02952816ad6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.049027] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6193b447-6ede-4dd7-ba95-d94ec3b5098e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.060096] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb5f565-c22f-44b3-8e85-0814eb0853e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.064101] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266978, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675735} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.064396] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. [ 916.065490] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aafc42c-ef93-4252-8a2a-05fbbe91e3e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.080368] env[62558]: DEBUG nova.compute.provider_tree [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.102958] env[62558]: DEBUG nova.objects.base [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Object Instance<63050323-047e-4d73-91ae-859467b4b5a7> lazy-loaded attributes: info_cache,migration_context {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 916.110836] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.112376] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd23766-69d2-455f-ac1e-d98293c1d1a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.115629] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ad89d6a-6850-4763-abbb-e150b3940fbf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.151606] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec67a34-e1a8-429d-99b3-9ef73aed2d98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.154500] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 916.154500] env[62558]: value = "task-1266980" [ 916.154500] env[62558]: _type = "Task" [ 916.154500] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.156163] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.156468] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.167525] env[62558]: DEBUG oslo_vmware.api [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 916.167525] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ae1e44-dd80-2c45-2bab-9c5970088f4a" [ 916.167525] env[62558]: _type = "Task" [ 916.167525] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.175902] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.181423] env[62558]: DEBUG oslo_vmware.api [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ae1e44-dd80-2c45-2bab-9c5970088f4a, 'name': SearchDatastore_Task, 'duration_secs': 0.010083} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.181704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.384134] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.475588] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266979, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072642} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.475876] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.476671] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6e366c-b2c4-40dd-bb87-6a3d3874dfac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.496636] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] f642911c-bb9e-4187-8e01-65c45cb6e793/f642911c-bb9e-4187-8e01-65c45cb6e793.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.496937] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec7950aa-fed7-4d19-987a-0e4fae6e1537 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.518055] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 916.518055] env[62558]: value = "task-1266981" [ 916.518055] env[62558]: _type = "Task" [ 916.518055] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.526722] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.584297] env[62558]: DEBUG nova.scheduler.client.report [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.667670] env[62558]: DEBUG nova.compute.utils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.668164] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.671343] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.700136] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.700413] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.700609] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.700834] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.700996] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.701167] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.701381] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.701543] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.701715] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.701892] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.702079] env[62558]: DEBUG nova.virt.hardware [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.703190] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d936653-5794-4e79-bee9-82048ef61bab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.713028] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a2b5d2-365d-40dc-ab98-b815d0ec7d1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.028510] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266981, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.093021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.093021] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.094469] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.532s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.094751] env[62558]: DEBUG nova.objects.instance [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lazy-loading 'pci_requests' on Instance uuid 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.124035] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Successfully updated port: 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.130990] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.130990] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.130990] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.131342] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.131566] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.134489] env[62558]: INFO nova.compute.manager [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Terminating instance [ 917.136730] env[62558]: DEBUG nova.compute.manager [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.136989] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.138111] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9803c166-8e10-4312-b2ae-a17b7719ef2c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.149783] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.151099] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2ca0226-7e7e-4799-94b8-2e633ba8838b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.167714] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.173790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.017s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.226297] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.226549] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.226713] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleting the datastore file [datastore2] ecf01ab6-5019-4453-b102-0e754abc4ef8 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.226989] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76502e11-9962-4e99-a9cd-f47ea01ee081 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.233523] env[62558]: DEBUG oslo_vmware.api [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 917.233523] env[62558]: value = "task-1266983" [ 917.233523] env[62558]: _type = "Task" [ 917.233523] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.243211] env[62558]: DEBUG oslo_vmware.api [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.345954] env[62558]: DEBUG nova.compute.manager [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Received event network-vif-plugged-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.346223] env[62558]: DEBUG oslo_concurrency.lockutils [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.346460] env[62558]: DEBUG oslo_concurrency.lockutils [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.346635] env[62558]: DEBUG oslo_concurrency.lockutils [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.347344] env[62558]: DEBUG nova.compute.manager [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] No waiting events found dispatching network-vif-plugged-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.347344] env[62558]: WARNING nova.compute.manager [req-f58903eb-d730-4603-9181-07457c08cb9d req-b6c1eda9-ec0d-481b-a634-32b3e56e9c69 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Received unexpected event network-vif-plugged-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a for instance with vm_state building and task_state spawning. [ 917.529312] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266981, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.597940] env[62558]: DEBUG nova.compute.utils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.600737] env[62558]: DEBUG nova.objects.instance [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lazy-loading 'numa_topology' on Instance uuid 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.601853] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Not allocating networking since 'none' was specified. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 917.628848] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.628990] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.629153] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.666021] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266980, 'name': ReconfigVM_Task, 'duration_secs': 1.303358} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.666376] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.667312] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9dc2b3-e26a-4ab3-a42b-7b60679685a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.693192] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8d2a0ed-8cf1-41c8-9d3e-d638656480fc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.713379] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 917.713379] env[62558]: value = "task-1266984" [ 917.713379] env[62558]: _type = "Task" [ 917.713379] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.724685] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266984, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.745149] env[62558]: DEBUG oslo_vmware.api [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1266983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141617} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.745447] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.745639] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.745819] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.745994] env[62558]: INFO nova.compute.manager [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Took 0.61 seconds to destroy the instance on the hypervisor. [ 917.746259] env[62558]: DEBUG oslo.service.loopingcall [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.746451] env[62558]: DEBUG nova.compute.manager [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.746541] env[62558]: DEBUG nova.network.neutron [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.029928] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266981, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.103598] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.106488] env[62558]: INFO nova.compute.claims [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.163575] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.214780] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.215150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.216023] env[62558]: INFO nova.compute.manager [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Attaching volume c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78 to /dev/sdb [ 918.230741] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266984, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.256057] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b8ea07-3032-429d-b009-98608c7b914f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.264068] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ff3918-5288-4b88-9de2-e4259d79bd0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.280553] env[62558]: DEBUG nova.virt.block_device [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating existing volume attachment record: a4e25e84-3178-4adb-8f74-61d5c958f031 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 918.332776] env[62558]: DEBUG nova.network.neutron [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating instance_info_cache with network_info: [{"id": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "address": "fa:16:3e:a0:70:5f", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff73e0e-0c", "ovs_interfaceid": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.483794] env[62558]: DEBUG nova.network.neutron [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.529847] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266981, 'name': ReconfigVM_Task, 'duration_secs': 1.544269} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.530170] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Reconfigured VM instance instance-00000055 to attach disk [datastore2] f642911c-bb9e-4187-8e01-65c45cb6e793/f642911c-bb9e-4187-8e01-65c45cb6e793.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.534604] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-828e205b-c9c0-44ec-ba48-3054372b65cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.537509] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 918.537509] env[62558]: value = "task-1266987" [ 918.537509] env[62558]: _type = "Task" [ 918.537509] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.545289] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266987, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.734197] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266984, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.834986] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.835418] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Instance network_info: |[{"id": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "address": "fa:16:3e:a0:70:5f", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff73e0e-0c", "ovs_interfaceid": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.835975] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:70:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.845744] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating folder: Project (46aa5789fccf4925aa49f9c4050c8463). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.848794] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e5bc694-2f8d-4a63-8904-be2c4a2433ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.865963] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created folder: Project (46aa5789fccf4925aa49f9c4050c8463) in parent group-v272451. [ 918.866345] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating folder: Instances. Parent ref: group-v272573. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.866689] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15ac2da0-99e6-46bd-9908-1ce7f24fa557 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.880562] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created folder: Instances in parent group-v272573. [ 918.880991] env[62558]: DEBUG oslo.service.loopingcall [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.882289] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.882994] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc84fea3-eab1-4333-93ff-1ad75174dfa6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.908859] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.908859] env[62558]: value = "task-1266991" [ 918.908859] env[62558]: _type = "Task" [ 918.908859] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.923200] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266991, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.928299] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece68cff-b8be-467b-a58e-01f48911dd77 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.936327] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16042187-a048-488b-b4b3-5ce281b5a9bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.976833] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91c20c3-8717-4fff-bc25-5a3ea32362cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.985648] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65612a2e-5414-4225-9f22-3933be890810 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.989872] env[62558]: INFO nova.compute.manager [-] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Took 1.24 seconds to deallocate network for instance. [ 919.005936] env[62558]: DEBUG nova.compute.provider_tree [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.047516] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266987, 'name': Rename_Task, 'duration_secs': 0.368763} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.047811] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.048082] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70d41b0a-9ba1-46a6-8585-c2cf02d60ed2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.055651] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 919.055651] env[62558]: value = "task-1266992" [ 919.055651] env[62558]: _type = "Task" [ 919.055651] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.064270] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.119823] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.151996] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.152314] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.152506] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.152741] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.152926] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.153117] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.153362] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.153531] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.153730] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.153963] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.154229] env[62558]: DEBUG nova.virt.hardware [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.155289] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71b1adc-b6bf-448a-b1d6-be711be57858 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.166349] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef92e35-1f54-489b-8bca-c2ce79849c92 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.183595] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.189653] env[62558]: DEBUG oslo.service.loopingcall [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.189955] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.190206] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44bebfc2-271d-44fa-a26e-19ae3452d61e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.208268] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.208268] env[62558]: value = "task-1266993" [ 919.208268] env[62558]: _type = "Task" [ 919.208268] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.217643] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266993, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.226797] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266984, 'name': ReconfigVM_Task, 'duration_secs': 1.180616} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.227149] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.227428] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95d740f8-63c9-4034-88c3-2c9787872b65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.235472] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 919.235472] env[62558]: value = "task-1266994" [ 919.235472] env[62558]: _type = "Task" [ 919.235472] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.244664] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.378552] env[62558]: DEBUG nova.compute.manager [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Received event network-changed-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.379046] env[62558]: DEBUG nova.compute.manager [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Refreshing instance network info cache due to event network-changed-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.379101] env[62558]: DEBUG oslo_concurrency.lockutils [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] Acquiring lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.379278] env[62558]: DEBUG oslo_concurrency.lockutils [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] Acquired lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.379455] env[62558]: DEBUG nova.network.neutron [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Refreshing network info cache for port 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.420553] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266991, 'name': CreateVM_Task, 'duration_secs': 0.32879} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.420728] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.421462] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.421637] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.421956] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.422494] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47ae602c-d369-416e-9ecd-49e67b34b7c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.428576] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 919.428576] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b57509-9350-d11a-c6a4-b481715d7a92" [ 919.428576] env[62558]: _type = "Task" [ 919.428576] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.439111] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b57509-9350-d11a-c6a4-b481715d7a92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.497091] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.509350] env[62558]: DEBUG nova.scheduler.client.report [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.567472] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266992, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.720574] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1266993, 'name': CreateVM_Task, 'duration_secs': 0.287821} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.720574] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.720574] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.746531] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266994, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.940017] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b57509-9350-d11a-c6a4-b481715d7a92, 'name': SearchDatastore_Task, 'duration_secs': 0.010962} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.940017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.940308] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.941197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.941197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.941197] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.941438] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.941696] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.941921] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb73617e-8c97-4694-95ee-023d364bc382 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.944034] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-943f80bb-45f0-44ed-a129-808e3f7c75ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.951594] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 919.951594] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a4093-6294-da0d-ea68-6b6f4576fd1b" [ 919.951594] env[62558]: _type = "Task" [ 919.951594] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.956291] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.956483] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.957594] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccb821fc-e2db-41d0-b030-69791b50982f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.962753] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a4093-6294-da0d-ea68-6b6f4576fd1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.968733] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 919.968733] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb61e-0bdc-cfd7-b4d8-9e6e91c56e33" [ 919.968733] env[62558]: _type = "Task" [ 919.968733] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.978455] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb61e-0bdc-cfd7-b4d8-9e6e91c56e33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.015436] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.920s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.017770] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.745s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.020467] env[62558]: INFO nova.compute.claims [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.070610] env[62558]: DEBUG oslo_vmware.api [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266992, 'name': PowerOnVM_Task, 'duration_secs': 0.662769} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.070880] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.071099] env[62558]: INFO nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Took 7.89 seconds to spawn the instance on the hypervisor. [ 920.071289] env[62558]: DEBUG nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.072101] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca671d8-7442-4814-8d2d-3739d75ca65f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.094084] env[62558]: INFO nova.network.neutron [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 920.097475] env[62558]: DEBUG nova.network.neutron [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updated VIF entry in instance network info cache for port 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.099588] env[62558]: DEBUG nova.network.neutron [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating instance_info_cache with network_info: [{"id": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "address": "fa:16:3e:a0:70:5f", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff73e0e-0c", "ovs_interfaceid": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.247338] env[62558]: DEBUG oslo_vmware.api [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1266994, 'name': PowerOnVM_Task, 'duration_secs': 0.554866} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.247678] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.250868] env[62558]: DEBUG nova.compute.manager [None req-b702cbef-f855-4e7c-9250-f6ef2046ea1b tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.251606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-095c6e40-3e78-4616-bb76-3c5097281673 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.462922] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525a4093-6294-da0d-ea68-6b6f4576fd1b, 'name': SearchDatastore_Task, 'duration_secs': 0.020885} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.463139] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.463431] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.463777] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.482344] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527bb61e-0bdc-cfd7-b4d8-9e6e91c56e33, 'name': SearchDatastore_Task, 'duration_secs': 0.030159} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.483173] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62bf5918-f0bb-4826-b5e7-94204fcfb261 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.489501] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 920.489501] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4a25c-8b00-28dc-5dd9-093b7a0e0a4e" [ 920.489501] env[62558]: _type = "Task" [ 920.489501] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.497476] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4a25c-8b00-28dc-5dd9-093b7a0e0a4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.591799] env[62558]: INFO nova.compute.manager [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Took 19.79 seconds to build instance. [ 920.601308] env[62558]: DEBUG oslo_concurrency.lockutils [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] Releasing lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.601421] env[62558]: DEBUG nova.compute.manager [req-ffd770bd-6148-4e0e-b0bb-f84012e565b4 req-357b4232-c8af-426a-9c78-caf73879704f service nova] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Received event network-vif-deleted-f0d0d5f5-cb9e-499e-ad0c-3025f93b4655 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.000350] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4a25c-8b00-28dc-5dd9-093b7a0e0a4e, 'name': SearchDatastore_Task, 'duration_secs': 0.009032} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.000753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.001068] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a3f42034-3d2b-472a-89c0-5445cb6fb567/a3f42034-3d2b-472a-89c0-5445cb6fb567.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.001193] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.001367] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.001594] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d59ba57-0daa-45ff-a222-554d9dc20dff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.003732] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f49e36dd-39f2-453e-9522-b6e4ebc48d4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.011937] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 921.011937] env[62558]: value = "task-1266996" [ 921.011937] env[62558]: _type = "Task" [ 921.011937] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.013184] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.013367] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.017054] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01db0d96-8a9d-4957-9f67-2344d9be85d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.023259] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 921.023259] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fbd074-13c6-cd75-b122-f99215e9c826" [ 921.023259] env[62558]: _type = "Task" [ 921.023259] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.031776] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266996, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.042869] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fbd074-13c6-cd75-b122-f99215e9c826, 'name': SearchDatastore_Task, 'duration_secs': 0.008583} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.043969] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f57e8d-8533-4774-92bc-3030851d6d66 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.049809] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 921.049809] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5220e335-3a74-8241-1aeb-97561a7ffa16" [ 921.049809] env[62558]: _type = "Task" [ 921.049809] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.061418] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5220e335-3a74-8241-1aeb-97561a7ffa16, 'name': SearchDatastore_Task, 'duration_secs': 0.008881} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.061837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.061933] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.062271] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f437772-8063-4fa7-a2c8-46965c79c740 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.070186] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 921.070186] env[62558]: value = "task-1266997" [ 921.070186] env[62558]: _type = "Task" [ 921.070186] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.080614] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.095893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0976073d-0e34-4ada-b613-2f2a6bd167ed tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.306s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.307804] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2eb2ed-0d53-4f92-b555-951b3ae26610 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.318394] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb77c3db-cbc4-4976-b93a-08efefae5111 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.352044] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eb9cef-abc7-46f8-bf38-07cf27a02083 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.362200] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bed1b6-4565-4d21-9a3a-cfd804c82d3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.384240] env[62558]: DEBUG nova.compute.provider_tree [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.522880] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266996, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.581777] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266997, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.890469] env[62558]: DEBUG nova.scheduler.client.report [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.028975] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266996, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528774} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.031743] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a3f42034-3d2b-472a-89c0-5445cb6fb567/a3f42034-3d2b-472a-89c0-5445cb6fb567.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.032168] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.034092] env[62558]: DEBUG nova.compute.manager [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.034433] env[62558]: DEBUG oslo_concurrency.lockutils [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.034820] env[62558]: DEBUG oslo_concurrency.lockutils [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.035154] env[62558]: DEBUG oslo_concurrency.lockutils [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.035442] env[62558]: DEBUG nova.compute.manager [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] No waiting events found dispatching network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.035722] env[62558]: WARNING nova.compute.manager [req-eb840e27-0b85-4170-a366-3bc189424452 req-3bd87fd4-d41b-4312-b11e-13468d19c182 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received unexpected event network-vif-plugged-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b for instance with vm_state shelved_offloaded and task_state spawning. [ 922.036154] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-010559d1-8084-4256-b979-7472c68b1f37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.047842] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 922.047842] env[62558]: value = "task-1266998" [ 922.047842] env[62558]: _type = "Task" [ 922.047842] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.062138] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266998, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.063671] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.063937] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.064323] env[62558]: DEBUG nova.network.neutron [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.086491] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266997, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.130062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.130375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.130565] env[62558]: INFO nova.compute.manager [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Shelving [ 922.395991] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.396610] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.399231] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.632s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.399759] env[62558]: DEBUG nova.objects.instance [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 922.558257] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072757} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.558585] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.559316] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4a08fe-0a4b-4a33-b6a7-2cbca1cd4d7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.585273] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] a3f42034-3d2b-472a-89c0-5445cb6fb567/a3f42034-3d2b-472a-89c0-5445cb6fb567.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.588698] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e53914e-747b-46c6-bb2b-75563698efe1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.610654] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266997, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.611959] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 922.611959] env[62558]: value = "task-1266999" [ 922.611959] env[62558]: _type = "Task" [ 922.611959] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.621484] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.639301] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.639627] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a87c0f09-cbf9-4997-b3eb-6b37fb67f0fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.646507] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 922.646507] env[62558]: value = "task-1267000" [ 922.646507] env[62558]: _type = "Task" [ 922.646507] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.655302] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.829271] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 922.829521] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272572', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'name': 'volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4815ba3f-265f-466a-9850-4c325cdb88de', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'serial': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 922.830402] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a3c961-7b6e-4638-8259-1e378c786dce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.852030] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29afeae6-9617-4236-998c-c7c43db01a91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.872585] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78/volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.872893] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ee121b1-9b8c-4535-9ea0-aa4a817aceb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.892972] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 922.892972] env[62558]: value = "task-1267001" [ 922.892972] env[62558]: _type = "Task" [ 922.892972] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.901198] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.909111] env[62558]: DEBUG nova.compute.utils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.910541] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.910843] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.958347] env[62558]: DEBUG nova.policy [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61cb5cd30d434f1eb594dd4521dab2d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e1d3e1dcd64e918c1e436e691e290b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.031517] env[62558]: DEBUG nova.network.neutron [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.087956] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1266997, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.839572} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.088455] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.088852] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.089250] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df1843b1-cc8f-4a5c-95b7-e0a5709c28da {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.096598] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 923.096598] env[62558]: value = "task-1267002" [ 923.096598] env[62558]: _type = "Task" [ 923.096598] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.105688] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.122370] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266999, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.156260] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267000, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.248425] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Successfully created port: d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.404478] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267001, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.411939] env[62558]: DEBUG oslo_concurrency.lockutils [None req-da1fb7e0-4527-49ac-b083-a0e79cfed25e tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.414067] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.816s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.418508] env[62558]: INFO nova.compute.claims [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.418508] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.534989] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='26dfcf32288b8bf0708876127456357e',container_format='bare',created_at=2024-10-11T05:05:54Z,direct_url=,disk_format='vmdk',id=2f0f0be9-1e05-4998-a402-10590e152a59,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2012413113-shelved',owner='480afedc5a8845e8920815407f4485dc',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-11T05:06:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.560474] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.560837] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.560837] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.560995] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.562110] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.562110] env[62558]: DEBUG nova.virt.hardware [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.562265] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73cf3dd-1387-4e4b-81bb-f852e17e0772 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.570475] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5523fd-af45-4639-9d7a-177b3f62b2f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.584402] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:c8:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7472a75c-48b2-4f52-a2b5-905ae7a9ef3b', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.592110] env[62558]: DEBUG oslo.service.loopingcall [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.592676] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.592935] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7e5738b-dc82-4daa-98fc-1c23d94dd2ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.615783] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071375} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.619579] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.619833] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.619833] env[62558]: value = "task-1267003" [ 923.619833] env[62558]: _type = "Task" [ 923.619833] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.620495] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3981e17-add7-4c47-a430-9976170a8dc8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.629308] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1266999, 'name': ReconfigVM_Task, 'duration_secs': 0.609647} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.629868] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfigured VM instance instance-00000056 to attach disk [datastore2] a3f42034-3d2b-472a-89c0-5445cb6fb567/a3f42034-3d2b-472a-89c0-5445cb6fb567.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.630479] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60e1cae8-fae8-40a3-9a0e-7e9e30a71cf0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.650151] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.650399] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267003, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.651125] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec05b5d7-905d-425c-a7cd-e4403536ed16 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.670601] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 923.670601] env[62558]: value = "task-1267004" [ 923.670601] env[62558]: _type = "Task" [ 923.670601] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.678098] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267000, 'name': PowerOffVM_Task, 'duration_secs': 0.588232} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.679472] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.679832] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 923.679832] env[62558]: value = "task-1267005" [ 923.679832] env[62558]: _type = "Task" [ 923.679832] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.680552] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4861e4-ac2e-42d9-ad0d-5e079b67af2c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.689624] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267004, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.695874] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267005, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.710075] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29d7d25-30d9-430e-a7a1-e5df7e2c6469 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.903363] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267001, 'name': ReconfigVM_Task, 'duration_secs': 0.937172} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.903688] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78/volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.908519] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf676c9d-a620-45b5-9f30-8ce77dece0b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.931476] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 923.931476] env[62558]: value = "task-1267006" [ 923.931476] env[62558]: _type = "Task" [ 923.931476] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.942989] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.067906] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.068276] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing instance network info cache due to event network-changed-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.068538] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquiring lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.068681] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquired lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.068852] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Refreshing network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.132857] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267003, 'name': CreateVM_Task, 'duration_secs': 0.305896} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.133071] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.133810] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.134190] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.134443] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.134723] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbbce725-9ad9-499c-b284-2482479ff553 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.139614] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 924.139614] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521f7971-b813-f5e2-f0a7-d103ad02911d" [ 924.139614] env[62558]: _type = "Task" [ 924.139614] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.148408] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521f7971-b813-f5e2-f0a7-d103ad02911d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.180610] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267004, 'name': Rename_Task, 'duration_secs': 0.191135} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.180913] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.181186] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f52b3dc-c43f-4831-a331-a2c81cd9057d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.187908] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 924.187908] env[62558]: value = "task-1267007" [ 924.187908] env[62558]: _type = "Task" [ 924.187908] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.194246] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267005, 'name': ReconfigVM_Task, 'duration_secs': 0.348612} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.194816] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Reconfigured VM instance instance-00000057 to attach disk [datastore2] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.195482] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a993c147-71aa-4361-b758-fccb0bf537b1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.199732] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.204540] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 924.204540] env[62558]: value = "task-1267008" [ 924.204540] env[62558]: _type = "Task" [ 924.204540] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.212420] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267008, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.220526] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 924.220790] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f7625f2c-fac4-4848-9262-1c7a152a6015 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.229434] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 924.229434] env[62558]: value = "task-1267009" [ 924.229434] env[62558]: _type = "Task" [ 924.229434] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.239594] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267009, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.429243] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.442911] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267006, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.457690] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1c493ebdbdae9636c6a70fad68235fc6',container_format='bare',created_at=2024-10-11T05:06:19Z,direct_url=,disk_format='vmdk',id=fe8845c5-445d-4f71-8ce6-a1a0ce770a67,min_disk=1,min_ram=0,name='tempest-test-snap-491702292',owner='52e1d3e1dcd64e918c1e436e691e290b',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-11T05:06:34Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.459018] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.459327] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.459327] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.459451] env[62558]: DEBUG nova.virt.hardware [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.460254] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3e7a2a-b9d5-4b18-af3e-1038c74dba80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.469683] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72825e5-884a-42e3-b639-bf9f42b6d965 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.652144] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.652431] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Processing image 2f0f0be9-1e05-4998-a402-10590e152a59 {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.652695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.652848] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.653032] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.653299] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fdb6459-8c32-4abd-9218-4fab59c8ddcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.666223] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.666420] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.669171] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fcc288f-f136-46ba-9748-df86c74b7da7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.674938] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 924.674938] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52874de3-69fe-7758-b779-d43ba27a919a" [ 924.674938] env[62558]: _type = "Task" [ 924.674938] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.684818] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52874de3-69fe-7758-b779-d43ba27a919a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.697733] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.714752] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267008, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.725929] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Successfully updated port: d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.732598] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed125a6b-31eb-4bc5-be77-f412ebad1de3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.752076] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267009, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.753566] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cbfd32-d1a7-4a4a-bbdf-013580b490c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.784823] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8103c1b3-66ce-4f83-838c-94e6d5592b1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.795063] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde574bc-951b-4050-b3e1-a110878f4747 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.811295] env[62558]: DEBUG nova.compute.provider_tree [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.946370] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267006, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.055771] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updated VIF entry in instance network info cache for port 7472a75c-48b2-4f52-a2b5-905ae7a9ef3b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 925.055771] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [{"id": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "address": "fa:16:3e:63:c8:49", "network": {"id": "c3ddffb4-b5d9-4edd-8cb2-a48b31d87ad5", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1163988664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "480afedc5a8845e8920815407f4485dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7472a75c-48", "ovs_interfaceid": "7472a75c-48b2-4f52-a2b5-905ae7a9ef3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.186858] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Preparing fetch location {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 925.187154] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Fetch image to [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39/OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39.vmdk {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 925.187346] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Downloading stream optimized image 2f0f0be9-1e05-4998-a402-10590e152a59 to [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39/OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39.vmdk on the data store datastore2 as vApp {{(pid=62558) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 925.187520] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Downloading image file data 2f0f0be9-1e05-4998-a402-10590e152a59 to the ESX as VM named 'OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39' {{(pid=62558) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 925.195270] env[62558]: DEBUG nova.compute.manager [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Received event network-vif-plugged-d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.195522] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] Acquiring lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.195752] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.196055] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.196256] env[62558]: DEBUG nova.compute.manager [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] No waiting events found dispatching network-vif-plugged-d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.196460] env[62558]: WARNING nova.compute.manager [req-ecfba8c2-543f-4a15-a403-5e5033b12bcd req-5104a618-fcdf-4815-a9aa-4d351bec1a86 service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Received unexpected event network-vif-plugged-d08a37d6-d27f-467e-9138-f2ff9de7f070 for instance with vm_state building and task_state spawning. [ 925.202160] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.230747] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.230912] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.231057] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.236328] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267008, 'name': Rename_Task, 'duration_secs': 0.903256} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.240220] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.241133] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c20351d-fe6d-4fe7-b57f-ff7b17ee35ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.248385] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267009, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.249772] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 925.249772] env[62558]: value = "task-1267010" [ 925.249772] env[62558]: _type = "Task" [ 925.249772] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.260550] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.278258] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 925.278258] env[62558]: value = "resgroup-9" [ 925.278258] env[62558]: _type = "ResourcePool" [ 925.278258] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 925.278258] env[62558]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-06c395a2-70ec-4780-a677-e73b9887ced4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.300077] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lease: (returnval){ [ 925.300077] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520d56d0-8f27-8d71-518f-d66fb4f84474" [ 925.300077] env[62558]: _type = "HttpNfcLease" [ 925.300077] env[62558]: } obtained for vApp import into resource pool (val){ [ 925.300077] env[62558]: value = "resgroup-9" [ 925.300077] env[62558]: _type = "ResourcePool" [ 925.300077] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 925.300348] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the lease: (returnval){ [ 925.300348] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520d56d0-8f27-8d71-518f-d66fb4f84474" [ 925.300348] env[62558]: _type = "HttpNfcLease" [ 925.300348] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 925.306966] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.306966] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520d56d0-8f27-8d71-518f-d66fb4f84474" [ 925.306966] env[62558]: _type = "HttpNfcLease" [ 925.306966] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 925.314289] env[62558]: DEBUG nova.scheduler.client.report [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.445017] env[62558]: DEBUG oslo_vmware.api [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267006, 'name': ReconfigVM_Task, 'duration_secs': 1.150159} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.445409] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272572', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'name': 'volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4815ba3f-265f-466a-9850-4c325cdb88de', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'serial': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 925.558120] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Releasing lock "refresh_cache-8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.558435] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.558611] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing instance network info cache due to event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.558838] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.558991] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.559192] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.701064] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.747528] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267009, 'name': CreateSnapshot_Task, 'duration_secs': 1.203799} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.747831] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 925.749042] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d1a3b5-2d48-4fe4-b29e-d1b5c107e79f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.764818] env[62558]: DEBUG oslo_vmware.api [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267010, 'name': PowerOnVM_Task, 'duration_secs': 0.451708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.765364] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.765712] env[62558]: INFO nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Took 6.65 seconds to spawn the instance on the hypervisor. [ 925.765944] env[62558]: DEBUG nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.766708] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9136816-58ae-4388-b68e-f1fcb5d8e07a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.776881] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.811811] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.811811] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520d56d0-8f27-8d71-518f-d66fb4f84474" [ 925.811811] env[62558]: _type = "HttpNfcLease" [ 925.811811] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 925.812373] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 925.812373] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520d56d0-8f27-8d71-518f-d66fb4f84474" [ 925.812373] env[62558]: _type = "HttpNfcLease" [ 925.812373] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 925.813314] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a107e4e-0573-49e6-b213-671daf09219e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.822270] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.822746] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.825729] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 925.825963] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk. {{(pid=62558) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 925.827655] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.954s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.829243] env[62558]: INFO nova.compute.claims [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.900728] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae73d73d-0c63-4910-9256-4fd74ae95cc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.110299] env[62558]: DEBUG nova.network.neutron [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Updating instance_info_cache with network_info: [{"id": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "address": "fa:16:3e:17:9b:b8", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd08a37d6-d2", "ovs_interfaceid": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.202592] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.252604] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updated VIF entry in instance network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.253103] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.273833] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 926.275048] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fe3af279-0993-4ae7-98ae-8a07da8ab294 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.287903] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 926.287903] env[62558]: value = "task-1267012" [ 926.287903] env[62558]: _type = "Task" [ 926.287903] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.293988] env[62558]: INFO nova.compute.manager [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Took 25.02 seconds to build instance. [ 926.301791] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267012, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.329233] env[62558]: DEBUG nova.compute.utils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.334021] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.334021] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 926.377928] env[62558]: DEBUG nova.policy [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43c75ab2a7474032b7084a0b58442d60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd567aaa721d04cac8a80a34c7104fb05', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.437936] env[62558]: DEBUG nova.compute.manager [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.438197] env[62558]: DEBUG nova.compute.manager [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing instance network info cache due to event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.438458] env[62558]: DEBUG oslo_concurrency.lockutils [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.496134] env[62558]: DEBUG nova.objects.instance [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'flavor' on Instance uuid 4815ba3f-265f-466a-9850-4c325cdb88de {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.613525] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.613858] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Instance network_info: |[{"id": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "address": "fa:16:3e:17:9b:b8", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd08a37d6-d2", "ovs_interfaceid": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.614344] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:9b:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd08a37d6-d27f-467e-9138-f2ff9de7f070', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.623162] env[62558]: DEBUG oslo.service.loopingcall [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.630011] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.630011] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-777a803e-570e-4aab-b98c-8374e2dc3c87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.657483] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.657483] env[62558]: value = "task-1267013" [ 926.657483] env[62558]: _type = "Task" [ 926.657483] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.670881] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267013, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.701971] env[62558]: DEBUG oslo_vmware.api [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267007, 'name': PowerOnVM_Task, 'duration_secs': 2.475688} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.704115] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.704336] env[62558]: INFO nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Took 10.03 seconds to spawn the instance on the hypervisor. [ 926.704481] env[62558]: DEBUG nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.705389] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1010071-df8c-41a0-8a74-1d50c87a3276 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.731565] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Successfully created port: 1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.758026] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.758026] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.758026] env[62558]: DEBUG nova.compute.manager [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing instance network info cache due to event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.758026] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.758026] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.758359] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.796430] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fd0aa911-0742-42b6-938a-8766a5ef00d0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.537s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.802645] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267012, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.834738] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.847421] env[62558]: INFO nova.compute.manager [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Rebuilding instance [ 926.902608] env[62558]: DEBUG nova.compute.manager [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.903641] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9016ae7-4e32-433f-a0bc-3495a6636456 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.956989] env[62558]: INFO nova.compute.manager [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Rescuing [ 926.957370] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.957678] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.957815] env[62558]: DEBUG nova.network.neutron [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.004497] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4adfe999-cd7b-4532-baf1-8cc07666e0b0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.789s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.177274] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267013, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.227334] env[62558]: INFO nova.compute.manager [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Took 26.14 seconds to build instance. [ 927.231560] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcc5721-512a-43ae-a74f-99386adb8a93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.243095] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b9592a-a2d4-4628-99b5-f7fc7450d6ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.290223] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf0faca-ed6d-4fc2-bb50-7155536e438e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.309749] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579a7ce1-1407-4560-b265-57ed019074d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.311806] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267012, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.327733] env[62558]: DEBUG nova.compute.provider_tree [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.333909] env[62558]: DEBUG nova.compute.manager [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Received event network-changed-d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.333909] env[62558]: DEBUG nova.compute.manager [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Refreshing instance network info cache due to event network-changed-d08a37d6-d27f-467e-9138-f2ff9de7f070. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.333909] env[62558]: DEBUG oslo_concurrency.lockutils [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] Acquiring lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.334156] env[62558]: DEBUG oslo_concurrency.lockutils [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] Acquired lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.334193] env[62558]: DEBUG nova.network.neutron [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Refreshing network info cache for port d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.340570] env[62558]: INFO nova.virt.block_device [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Booting with volume 5dc54ac6-cb87-4624-a362-73c54c37b21e at /dev/sda [ 927.387364] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4a8c2b9-d81b-493c-be07-47ac74eed7fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.402361] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcdf6fb-2fa2-4a03-9329-2554169ff167 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.419884] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.420295] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b468e0ca-5e48-4b44-9f6d-0fd0a1496aa2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.444038] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 927.444038] env[62558]: value = "task-1267014" [ 927.444038] env[62558]: _type = "Task" [ 927.444038] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.447382] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e997ebb-75cc-4373-828e-f8889c6242ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.460385] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.466326] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4299f503-bbff-4a0b-b3d2-3c80be762ce2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.506839] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79e5ae5-eda9-41d2-bc52-7bfc26ece17e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.515137] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a1949d-b48c-475d-908f-c9e0b6c329ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.533787] env[62558]: DEBUG nova.virt.block_device [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating existing volume attachment record: d184c5e3-482d-43c2-a27d-9cce4e3be0a0 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 927.675567] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267013, 'name': CreateVM_Task, 'duration_secs': 0.890836} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.675567] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.676201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.677450] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.677450] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.678703] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf7f46ab-c12d-434a-b8e8-3ef88e8dab3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.684346] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 927.684346] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521d3965-a873-68d1-23f6-433994250702" [ 927.684346] env[62558]: _type = "Task" [ 927.684346] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.689270] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Completed reading data from the image iterator. {{(pid=62558) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 927.689472] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 927.690286] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c708fac-1b2c-409c-8f0f-e78cad1409ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.699916] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521d3965-a873-68d1-23f6-433994250702, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.701400] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 927.701569] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 927.701833] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-93d07931-2e42-42c5-a6c4-2496719a708c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.730385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3178e56a-f98d-4e54-996f-a2030ccc8f71 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.654s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.804915] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267012, 'name': CloneVM_Task, 'duration_secs': 1.394105} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.805152] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Created linked-clone VM from snapshot [ 927.805891] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f415673-20ae-4b5a-b7d1-f49ab8f0d295 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.813650] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Uploading image bf57715a-d9cb-4e6c-885f-8cbb16b14efd {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 927.837627] env[62558]: DEBUG nova.scheduler.client.report [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.842483] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 927.842483] env[62558]: value = "vm-272580" [ 927.842483] env[62558]: _type = "VirtualMachine" [ 927.842483] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 927.844598] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-59c28da0-7525-4f08-9792-2ee8b2385629 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.855591] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lease: (returnval){ [ 927.855591] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523fdf1f-b6c4-f6bb-b767-27eca3153d04" [ 927.855591] env[62558]: _type = "HttpNfcLease" [ 927.855591] env[62558]: } obtained for exporting VM: (result){ [ 927.855591] env[62558]: value = "vm-272580" [ 927.855591] env[62558]: _type = "VirtualMachine" [ 927.855591] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 927.855933] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the lease: (returnval){ [ 927.855933] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523fdf1f-b6c4-f6bb-b767-27eca3153d04" [ 927.855933] env[62558]: _type = "HttpNfcLease" [ 927.855933] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 927.862609] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.862609] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523fdf1f-b6c4-f6bb-b767-27eca3153d04" [ 927.862609] env[62558]: _type = "HttpNfcLease" [ 927.862609] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 927.880291] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updated VIF entry in instance network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.880291] env[62558]: DEBUG nova.network.neutron [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.962661] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267014, 'name': PowerOffVM_Task, 'duration_secs': 0.164942} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.962999] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.963261] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.964090] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856d4af7-0a08-43d6-9435-eca9b322b46f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.971942] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.972214] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ced76cbc-2fc9-4748-bb50-c8ac57ed2fef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.000353] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.000630] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.000876] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleting the datastore file [datastore2] f8d6df86-52b8-4692-add4-1ffec82cc598 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.001204] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4e6e116-fff0-4725-9d44-b5f50881e42b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.008993] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 928.008993] env[62558]: value = "task-1267017" [ 928.008993] env[62558]: _type = "Task" [ 928.008993] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.017345] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.085778] env[62558]: DEBUG nova.network.neutron [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Updated VIF entry in instance network info cache for port d08a37d6-d27f-467e-9138-f2ff9de7f070. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.086194] env[62558]: DEBUG nova.network.neutron [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Updating instance_info_cache with network_info: [{"id": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "address": "fa:16:3e:17:9b:b8", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd08a37d6-d2", "ovs_interfaceid": "d08a37d6-d27f-467e-9138-f2ff9de7f070", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.095679] env[62558]: DEBUG oslo_vmware.rw_handles [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c6f9cf-86e2-f067-a91a-d86ca19f97e8/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 928.095679] env[62558]: INFO nova.virt.vmwareapi.images [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Downloaded image file data 2f0f0be9-1e05-4998-a402-10590e152a59 [ 928.096852] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc55b61b-97ce-4e2e-8fc0-0a2dd5900687 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.114875] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f544b52d-73e2-4a71-9a63-7503c494c518 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.194808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.195108] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Processing image fe8845c5-445d-4f71-8ce6-a1a0ce770a67 {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.195356] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.195507] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.195685] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.196085] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de3c2368-cdb3-4c04-abe9-e169f2644362 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.207058] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.207258] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.207973] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc900b4c-8489-45ad-a46b-da1b8cf87f1c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.214203] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 928.214203] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fab04c-98c7-16b1-98f0-e657e6a75969" [ 928.214203] env[62558]: _type = "Task" [ 928.214203] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.222241] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fab04c-98c7-16b1-98f0-e657e6a75969, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.269269] env[62558]: DEBUG nova.network.neutron [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.286766] env[62558]: INFO nova.virt.vmwareapi.images [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] The imported VM was unregistered [ 928.289193] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Caching image {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 928.289430] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59 {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.289988] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbc9154b-b0fe-4656-8fb8-bfc8bda7c339 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.300213] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Successfully updated port: 1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.310673] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Created directory with path [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59 {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.310872] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39/OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39.vmdk to [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk. {{(pid=62558) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 928.311153] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-11540517-47be-4cca-8272-53541d41df5e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.321445] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 928.321445] env[62558]: value = "task-1267019" [ 928.321445] env[62558]: _type = "Task" [ 928.321445] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.330797] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.350026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.350026] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.352638] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.681s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.352853] env[62558]: DEBUG nova.objects.instance [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'resources' on Instance uuid 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.364824] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.364824] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523fdf1f-b6c4-f6bb-b767-27eca3153d04" [ 928.364824] env[62558]: _type = "HttpNfcLease" [ 928.364824] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 928.365177] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 928.365177] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523fdf1f-b6c4-f6bb-b767-27eca3153d04" [ 928.365177] env[62558]: _type = "HttpNfcLease" [ 928.365177] env[62558]: }. {{(pid=62558) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 928.365926] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb960f00-4c5f-4ff1-adf0-b4fe0bf49699 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.374130] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 928.374343] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk for reading. {{(pid=62558) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 928.433404] env[62558]: DEBUG oslo_concurrency.lockutils [req-235b8b68-2dce-4df4-91dc-cedaba730219 req-d2834781-19f6-4b35-a16d-6872c89408e1 service nova] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.436057] env[62558]: DEBUG oslo_concurrency.lockutils [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.436057] env[62558]: DEBUG nova.network.neutron [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.520933] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099476} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.521314] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.521383] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.521513] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.592124] env[62558]: DEBUG oslo_concurrency.lockutils [req-7f92624d-63e4-471e-8c54-a9d7438c42a6 req-2ea5a2df-b2cc-4049-b47b-6f5e5aa8762d service nova] Releasing lock "refresh_cache-42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.592124] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d8289bb-ad2b-4e45-9953-bb16aa490d89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.712017] env[62558]: DEBUG nova.compute.manager [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.712208] env[62558]: DEBUG nova.compute.manager [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing instance network info cache due to event network-changed-d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 928.712345] env[62558]: DEBUG oslo_concurrency.lockutils [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] Acquiring lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.725606] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Preparing fetch location {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 928.725856] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Fetch image to [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a/OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a.vmdk {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 928.726067] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Downloading stream optimized image fe8845c5-445d-4f71-8ce6-a1a0ce770a67 to [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a/OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a.vmdk on the data store datastore2 as vApp {{(pid=62558) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 928.726241] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Downloading image file data fe8845c5-445d-4f71-8ce6-a1a0ce770a67 to the ESX as VM named 'OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a' {{(pid=62558) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 928.773880] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.810251] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.810251] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquired lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.810251] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.819088] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 928.819088] env[62558]: value = "resgroup-9" [ 928.819088] env[62558]: _type = "ResourcePool" [ 928.819088] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 928.820598] env[62558]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-40f6c4d7-911b-4b44-923b-7656cd983db3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.865247] env[62558]: DEBUG nova.compute.utils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.865247] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.866259] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.866683] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 928.868658] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease: (returnval){ [ 928.868658] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 928.868658] env[62558]: _type = "HttpNfcLease" [ 928.868658] env[62558]: } obtained for vApp import into resource pool (val){ [ 928.868658] env[62558]: value = "resgroup-9" [ 928.868658] env[62558]: _type = "ResourcePool" [ 928.868658] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 928.869062] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the lease: (returnval){ [ 928.869062] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 928.869062] env[62558]: _type = "HttpNfcLease" [ 928.869062] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 928.877737] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.877737] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 928.877737] env[62558]: _type = "HttpNfcLease" [ 928.877737] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 928.944156] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.944466] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.944720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.944967] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.945170] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.949542] env[62558]: DEBUG nova.policy [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.954865] env[62558]: INFO nova.compute.manager [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Terminating instance [ 928.957473] env[62558]: DEBUG nova.compute.manager [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.957715] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 928.958961] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eba7e24-861e-4cbc-aa00-3fb289ec6380 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.970207] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.970620] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d29bdcd2-bf5f-4c14-8243-f180a273102e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.984481] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 928.984481] env[62558]: value = "task-1267021" [ 928.984481] env[62558]: _type = "Task" [ 928.984481] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.996347] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1267021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.260659] env[62558]: DEBUG nova.network.neutron [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updated VIF entry in instance network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.262690] env[62558]: DEBUG nova.network.neutron [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.300302] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d252f576-ae31-4166-a31d-523b6651335e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.314659] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3f5464-7e67-49a1-893c-74d458c7b3ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.364877] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eea884-3a5b-4ffa-a7fb-d2a64a098575 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.370221] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.374800] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.376405] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4981faa-e872-454b-8152-e59f51a1314d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.392794] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.395382] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.399225] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21437f03-7e03-46a4-9421-d8afaae74a3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.408313] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 929.408313] env[62558]: value = "task-1267022" [ 929.408313] env[62558]: _type = "Task" [ 929.408313] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.416422] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.416422] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 929.416422] env[62558]: _type = "HttpNfcLease" [ 929.416422] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 929.427632] env[62558]: DEBUG nova.compute.provider_tree [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.434671] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.505513] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1267021, 'name': PowerOffVM_Task, 'duration_secs': 0.423314} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.506335] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Successfully created port: 01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.508441] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.508704] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.509155] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b214567f-bd6f-4674-b134-d403c72a8eb9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.520294] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Received event network-vif-plugged-1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.520294] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Acquiring lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.520904] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.520904] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.521355] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] No waiting events found dispatching network-vif-plugged-1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.521355] env[62558]: WARNING nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Received unexpected event network-vif-plugged-1f94c71a-6387-41af-be7e-7fb674e8b3af for instance with vm_state building and task_state spawning. [ 929.521647] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Received event network-changed-1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.521831] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Refreshing instance network info cache due to event network-changed-1f94c71a-6387-41af-be7e-7fb674e8b3af. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.522215] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Acquiring lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.591032] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.591032] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.591032] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Deleting the datastore file [datastore1] 7aebdb8a-517b-4168-91e0-1b704e6a11c7 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.591350] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34fe2cc7-797b-4b92-906c-77bf8a0f72b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.602374] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for the task: (returnval){ [ 929.602374] env[62558]: value = "task-1267024" [ 929.602374] env[62558]: _type = "Task" [ 929.602374] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.616720] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1267024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.635502] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 929.636525] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.636823] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.636952] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.637176] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.637331] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.637669] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.638350] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.638632] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.638837] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.639693] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.639693] env[62558]: DEBUG nova.virt.hardware [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.641200] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1d34e3-1d7a-42e9-ade9-3fe994cbe69f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.654142] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7834f5dd-d60c-4f86-b3c4-fdad6f7940a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.751263] env[62558]: DEBUG nova.network.neutron [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating instance_info_cache with network_info: [{"id": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "address": "fa:16:3e:08:60:a5", "network": {"id": "adf89e17-7cea-47a6-8654-a97be0414a6c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1824548634-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d567aaa721d04cac8a80a34c7104fb05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f94c71a-63", "ovs_interfaceid": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.770562] env[62558]: DEBUG oslo_concurrency.lockutils [req-50a8ba21-238f-48d4-a897-6892784666cf req-9e895296-b4f4-4e2c-857b-158017c4d070 service nova] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.771443] env[62558]: DEBUG oslo_concurrency.lockutils [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] Acquired lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.771947] env[62558]: DEBUG nova.network.neutron [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Refreshing network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.868842] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.891996] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.891996] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 929.891996] env[62558]: _type = "HttpNfcLease" [ 929.891996] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 929.922459] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.938937] env[62558]: DEBUG nova.scheduler.client.report [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.118528] env[62558]: DEBUG oslo_vmware.api [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Task: {'id': task-1267024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.409047} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.119065] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.119410] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.119706] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.120010] env[62558]: INFO nova.compute.manager [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 930.120554] env[62558]: DEBUG oslo.service.loopingcall [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.121027] env[62558]: DEBUG nova.compute.manager [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.121293] env[62558]: DEBUG nova.network.neutron [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.254165] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Releasing lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.254532] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Instance network_info: |[{"id": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "address": "fa:16:3e:08:60:a5", "network": {"id": "adf89e17-7cea-47a6-8654-a97be0414a6c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1824548634-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d567aaa721d04cac8a80a34c7104fb05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f94c71a-63", "ovs_interfaceid": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.254882] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Acquired lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.255338] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Refreshing network info cache for port 1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.257574] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:60:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f94c71a-6387-41af-be7e-7fb674e8b3af', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.266825] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Creating folder: Project (d567aaa721d04cac8a80a34c7104fb05). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.266825] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf163ee4-c27b-41ed-813c-f88488edec9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.298820] env[62558]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 930.299403] env[62558]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62558) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 930.299841] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Folder already exists: Project (d567aaa721d04cac8a80a34c7104fb05). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 930.300068] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Creating folder: Instances. Parent ref: group-v272555. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.300337] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f9f39549-b334-4480-9094-25105451d4ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.316476] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Created folder: Instances in parent group-v272555. [ 930.316752] env[62558]: DEBUG oslo.service.loopingcall [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.317025] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.317562] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bbb35fc-a5f3-4dad-bed3-07d1d2dbf18e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.341600] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.341600] env[62558]: value = "task-1267027" [ 930.341600] env[62558]: _type = "Task" [ 930.341600] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.355198] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267027, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.368786] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.389112] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.391320] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 930.391320] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 930.391320] env[62558]: _type = "HttpNfcLease" [ 930.391320] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 930.394640] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 930.394640] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525086b3-1bef-31d8-0459-9447f976ebde" [ 930.394640] env[62558]: _type = "HttpNfcLease" [ 930.394640] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 930.395629] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f091a3d-9465-4f71-87f9-5480a675fb26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.406742] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 930.407217] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk. {{(pid=62558) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 930.479157] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.126s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.489176] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.537s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.490953] env[62558]: INFO nova.compute.claims [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.493830] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.500951] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-23b3ff7f-0a8a-4672-925e-4edf790146b1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.521161] env[62558]: INFO nova.scheduler.client.report [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocations for instance 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed [ 930.621758] env[62558]: DEBUG nova.network.neutron [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updated VIF entry in instance network info cache for port d6d6deff-eea7-4653-854b-69c6e43b962a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.622343] env[62558]: DEBUG nova.network.neutron [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [{"id": "d6d6deff-eea7-4653-854b-69c6e43b962a", "address": "fa:16:3e:45:3f:ad", "network": {"id": "aab0c431-04f1-4e99-999a-e34152c1ef69", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1077028713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a0182f4eec034883bfac1e6c85069265", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "535b175f-71d3-4226-81ab-ca253f27fedd", "external-id": "nsx-vlan-transportzone-155", "segmentation_id": 155, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d6deff-ee", "ovs_interfaceid": "d6d6deff-eea7-4653-854b-69c6e43b962a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.744644] env[62558]: DEBUG nova.compute.manager [req-2344d0a2-d035-44f5-a964-171a1507fbf6 req-ebab4e07-5f0a-4a3b-bf92-c5dab89cbec6 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Received event network-vif-deleted-d6d6deff-eea7-4653-854b-69c6e43b962a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.745073] env[62558]: INFO nova.compute.manager [req-2344d0a2-d035-44f5-a964-171a1507fbf6 req-ebab4e07-5f0a-4a3b-bf92-c5dab89cbec6 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Neutron deleted interface d6d6deff-eea7-4653-854b-69c6e43b962a; detaching it from the instance and deleting it from the info cache [ 930.745362] env[62558]: DEBUG nova.network.neutron [req-2344d0a2-d035-44f5-a964-171a1507fbf6 req-ebab4e07-5f0a-4a3b-bf92-c5dab89cbec6 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.855765] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267027, 'name': CreateVM_Task, 'duration_secs': 0.429145} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.857185] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.858848] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': 'd184c5e3-482d-43c2-a27d-9cce4e3be0a0', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272562', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'name': 'volume-5dc54ac6-cb87-4624-a362-73c54c37b21e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2480822a-e10a-4066-a5d8-5ca633ab9b12', 'attached_at': '', 'detached_at': '', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'serial': '5dc54ac6-cb87-4624-a362-73c54c37b21e'}, 'boot_index': 0, 'disk_bus': None, 'delete_on_termination': True, 'guest_format': None, 'mount_device': '/dev/sda', 'volume_type': None}], 'swap': None} {{(pid=62558) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 930.858848] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Root volume attach. Driver type: vmdk {{(pid=62558) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 930.859133] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb19c73f-27a4-4c2f-9d61-4314751de4a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.879347] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c99d90-3ef7-46e6-9e4f-e1064d09f4f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.882293] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.890603] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e750973c-db3e-4343-803f-8160504066df {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.908390] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-9b42952a-ca99-4eb8-9903-fa8a7e1bc536 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.917814] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.918840] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.919036] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.919302] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.919596] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.919751] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.920173] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.920572] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.920791] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.920974] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.921178] env[62558]: DEBUG nova.virt.hardware [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.928014] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd71d22-9c4e-4bdf-89b9-e8d94b7692d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.936483] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 930.936483] env[62558]: value = "task-1267028" [ 930.936483] env[62558]: _type = "Task" [ 930.936483] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.941346] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.941671] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.941957] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.942236] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.942455] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.942649] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.942903] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.943175] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.943387] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.943608] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.943797] env[62558]: DEBUG nova.virt.hardware [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.945762] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b363f8-e3b6-41dd-a62a-7aa42b2af0d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.964623] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687ec640-6e01-4517-9b30-c91f5d4c1b10 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.969681] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267022, 'name': PowerOffVM_Task, 'duration_secs': 1.266117} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.979245] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.980629] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9783ff-0415-4499-8aeb-d129b61b0174 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.985933] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc37852-1346-42c2-a211-d2a59e26e596 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.016035] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267028, 'name': RelocateVM_Task} progress is 17%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.016035] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance VIF info [] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.019211] env[62558]: DEBUG oslo.service.loopingcall [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.024985] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.026325] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f2418b4-8ee6-4994-9973-e9a522093823 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.078258] env[62558]: DEBUG nova.network.neutron [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.093349] env[62558]: DEBUG oslo_concurrency.lockutils [None req-89baf0dc-ddc5-430b-b759-d079f99fabed tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "6c2fedb1-6240-482d-9ad2-e8dabde4d7ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.355s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.098163] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab18b710-98aa-408d-93c0-b5834ff9a400 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.108410] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.108410] env[62558]: value = "task-1267029" [ 931.108410] env[62558]: _type = "Task" [ 931.108410] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.131542] env[62558]: DEBUG oslo_concurrency.lockutils [req-7b0bad8a-0d65-45de-a7a1-7f708c4260a3 req-259e97af-1729-45e8-a7d2-a37a29f41083 service nova] Releasing lock "refresh_cache-7aebdb8a-517b-4168-91e0-1b704e6a11c7" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.132496] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267029, 'name': CreateVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.146773] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.147638] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68849172-f461-4b2d-af3b-778b11f19058 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.159515] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 931.159515] env[62558]: value = "task-1267030" [ 931.159515] env[62558]: _type = "Task" [ 931.159515] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.173078] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 931.173463] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.173829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.174037] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.174372] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.174701] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db08b9aa-7a0e-4024-829d-d6b3c67a3532 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.182729] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.182969] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.186845] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e1c41e8-d9a0-4b39-98ec-ec16289f65a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.193324] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 931.193324] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522f80d4-c5a0-2fa6-f05f-f95105c47165" [ 931.193324] env[62558]: _type = "Task" [ 931.193324] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.206670] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522f80d4-c5a0-2fa6-f05f-f95105c47165, 'name': SearchDatastore_Task, 'duration_secs': 0.010445} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.210202] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cc99cc2-500b-4019-af84-8215aea54612 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.216191] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 931.216191] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523449a0-4cae-cc66-428e-715467e21f40" [ 931.216191] env[62558]: _type = "Task" [ 931.216191] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.229742] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523449a0-4cae-cc66-428e-715467e21f40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.248786] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15b4e61c-6aaa-431e-b238-eab6c43a94d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.259254] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebce9cf-83e2-42e1-8286-0f7a7095f4a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.299637] env[62558]: DEBUG nova.compute.manager [req-2344d0a2-d035-44f5-a964-171a1507fbf6 req-ebab4e07-5f0a-4a3b-bf92-c5dab89cbec6 service nova] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Detach interface failed, port_id=d6d6deff-eea7-4653-854b-69c6e43b962a, reason: Instance 7aebdb8a-517b-4168-91e0-1b704e6a11c7 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 931.368985] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267019, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.92135} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.369311] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39/OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39.vmdk to [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk. [ 931.369516] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Cleaning up location [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39 {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 931.369692] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_343a1434-b7ce-4c54-8ce3-ecd70e167b39 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.369972] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4daba3c-c1d3-4063-9faa-2c9d0e1477a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.377403] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 931.377403] env[62558]: value = "task-1267031" [ 931.377403] env[62558]: _type = "Task" [ 931.377403] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.385918] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.397940] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Completed reading data from the image iterator. {{(pid=62558) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 931.398312] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 931.399522] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689be9f2-fe25-4b00-a028-6f9ba961b505 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.406920] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 931.407222] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 931.407555] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-782dd500-ab19-4471-aa74-30d624616435 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.454934] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267028, 'name': RelocateVM_Task, 'duration_secs': 0.506029} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.455374] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 931.455508] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272562', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'name': 'volume-5dc54ac6-cb87-4624-a362-73c54c37b21e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2480822a-e10a-4066-a5d8-5ca633ab9b12', 'attached_at': '', 'detached_at': '', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'serial': '5dc54ac6-cb87-4624-a362-73c54c37b21e'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 931.456326] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb00a5d-32ec-44f0-944a-885f64ffd094 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.459685] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updated VIF entry in instance network info cache for port 1f94c71a-6387-41af-be7e-7fb674e8b3af. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.460052] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating instance_info_cache with network_info: [{"id": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "address": "fa:16:3e:08:60:a5", "network": {"id": "adf89e17-7cea-47a6-8654-a97be0414a6c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1824548634-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d567aaa721d04cac8a80a34c7104fb05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f94c71a-63", "ovs_interfaceid": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.475565] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7699baa3-fa20-4723-ac70-7de54a7ef7cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.499447] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] volume-5dc54ac6-cb87-4624-a362-73c54c37b21e/volume-5dc54ac6-cb87-4624-a362-73c54c37b21e.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.500394] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f8045aa-4e5d-4f17-9c82-74ca7d7a2e17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.516941] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Successfully updated port: 01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.527480] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 931.527480] env[62558]: value = "task-1267032" [ 931.527480] env[62558]: _type = "Task" [ 931.527480] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.538051] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267032, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.578828] env[62558]: DEBUG nova.compute.manager [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-plugged-01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.579144] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.579378] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.579643] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.579836] env[62558]: DEBUG nova.compute.manager [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] No waiting events found dispatching network-vif-plugged-01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.580023] env[62558]: WARNING nova.compute.manager [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received unexpected event network-vif-plugged-01b4d7b4-efd1-4cc4-aa28-2663beeb66be for instance with vm_state building and task_state spawning. [ 931.580169] env[62558]: DEBUG nova.compute.manager [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-changed-01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.580326] env[62558]: DEBUG nova.compute.manager [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing instance network info cache due to event network-changed-01b4d7b4-efd1-4cc4-aa28-2663beeb66be. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.580593] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.580805] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.581052] env[62558]: DEBUG nova.network.neutron [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing network info cache for port 01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.585830] env[62558]: INFO nova.compute.manager [-] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Took 1.46 seconds to deallocate network for instance. [ 931.622897] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267029, 'name': CreateVM_Task, 'duration_secs': 0.302248} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.623176] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.623669] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.624013] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.624423] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.624825] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91fac73b-585e-4c61-b8ce-4f8b1b154a1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.632769] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 931.632769] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521beb31-4fed-d77e-e7a7-e50b77004e30" [ 931.632769] env[62558]: _type = "Task" [ 931.632769] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.641703] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521beb31-4fed-d77e-e7a7-e50b77004e30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.645469] env[62558]: DEBUG oslo_vmware.rw_handles [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526f888b-d874-86d7-867a-5404ef37e05f/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 931.645694] env[62558]: INFO nova.virt.vmwareapi.images [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Downloaded image file data fe8845c5-445d-4f71-8ce6-a1a0ce770a67 [ 931.646719] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cf7e3a-88a4-463b-bda5-3b1b19f6b902 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.665103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4255932-2f1a-405f-bbd2-8be9ca3d9870 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.706367] env[62558]: INFO nova.virt.vmwareapi.images [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] The imported VM was unregistered [ 931.708075] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Caching image {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 931.708332] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating directory with path [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67 {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.708616] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a34a322-da9c-4018-a906-b32a8fc777cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.722345] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created directory with path [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67 {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.722575] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a/OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a.vmdk to [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk. {{(pid=62558) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 931.723277] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-93de01c8-bd1c-44fb-9bae-c8b62e851f14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.729290] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523449a0-4cae-cc66-428e-715467e21f40, 'name': SearchDatastore_Task, 'duration_secs': 0.010005} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.732320] env[62558]: DEBUG oslo_concurrency.lockutils [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.732567] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. {{(pid=62558) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 931.733176] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdfa7893-042a-4762-8c28-a37b351f7fcb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.736586] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 931.736586] env[62558]: value = "task-1267034" [ 931.736586] env[62558]: _type = "Task" [ 931.736586] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.741412] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 931.741412] env[62558]: value = "task-1267035" [ 931.741412] env[62558]: _type = "Task" [ 931.741412] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.748691] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.752671] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.855575] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cfd984b-2d2d-489d-b38d-a44fc62936bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.863718] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ce8d85-aa2f-4b35-8404-80baf2d7769c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.903472] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469f02ed-491a-450e-86a4-02ef7ccd8ccf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.912134] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035128} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.914641] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.914859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.915159] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk to [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.915543] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c9ea9ef-4095-4bcb-95a9-f147c402618b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.918751] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a326de-ffa0-4289-b432-e3ff8c623afe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.935378] env[62558]: DEBUG nova.compute.provider_tree [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.938246] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 931.938246] env[62558]: value = "task-1267036" [ 931.938246] env[62558]: _type = "Task" [ 931.938246] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.963809] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Releasing lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.964031] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Received event network-changed-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.964306] env[62558]: DEBUG nova.compute.manager [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Refreshing instance network info cache due to event network-changed-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.964470] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Acquiring lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.964651] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Acquired lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.964863] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Refreshing network info cache for port 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.023726] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.040723] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267032, 'name': ReconfigVM_Task, 'duration_secs': 0.328204} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.041149] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Reconfigured VM instance instance-00000059 to attach disk [datastore2] volume-5dc54ac6-cb87-4624-a362-73c54c37b21e/volume-5dc54ac6-cb87-4624-a362-73c54c37b21e.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.047220] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23fc1923-40c4-42c9-b709-c0d6be206515 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.067288] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 932.067288] env[62558]: value = "task-1267037" [ 932.067288] env[62558]: _type = "Task" [ 932.067288] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.075633] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267037, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.093180] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.127025] env[62558]: DEBUG nova.network.neutron [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.144622] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521beb31-4fed-d77e-e7a7-e50b77004e30, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.144965] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.145361] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.145664] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.145903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.146175] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.146488] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40e041b3-6d96-4ed2-a5c9-af009343b2a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.156433] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.156641] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.157464] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbd9b663-da5f-4ba5-a9e3-8623d28f895e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.163529] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 932.163529] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f744fa-d080-807c-e168-e02d4226decb" [ 932.163529] env[62558]: _type = "Task" [ 932.163529] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.172521] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f744fa-d080-807c-e168-e02d4226decb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.249217] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.255057] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267035, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.257137] env[62558]: DEBUG nova.network.neutron [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.439710] env[62558]: DEBUG nova.scheduler.client.report [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.454697] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.576431] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267037, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.676743] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f744fa-d080-807c-e168-e02d4226decb, 'name': SearchDatastore_Task, 'duration_secs': 0.017086} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.680047] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e68bf44-dafd-44fe-be3a-83348b2f59ec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.686308] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 932.686308] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52579ab4-a933-e2fe-bdf6-1b987181bce4" [ 932.686308] env[62558]: _type = "Task" [ 932.686308] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.695107] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52579ab4-a933-e2fe-bdf6-1b987181bce4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.749578] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.755285] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.952727} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.755605] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk. [ 932.756555] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c62b429-abad-43ac-8bf6-8c96a95a14c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.759548] env[62558]: DEBUG oslo_concurrency.lockutils [req-3cd56f7d-0259-4dd1-921f-f2cf5dfb6b41 req-38b963bd-69d3-4701-925c-df5405d53dc3 service nova] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.759920] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.760094] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.790050] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.791104] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26388a0c-d03f-4698-9e32-0ff2b942cbef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.814733] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 932.814733] env[62558]: value = "task-1267038" [ 932.814733] env[62558]: _type = "Task" [ 932.814733] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.825966] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.885218] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updated VIF entry in instance network info cache for port 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.885693] env[62558]: DEBUG nova.network.neutron [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating instance_info_cache with network_info: [{"id": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "address": "fa:16:3e:a0:70:5f", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff73e0e-0c", "ovs_interfaceid": "8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.950694] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.951522] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.958888] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 16.777s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.960668] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.079962] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267037, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.197598] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52579ab4-a933-e2fe-bdf6-1b987181bce4, 'name': SearchDatastore_Task, 'duration_secs': 0.019567} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.198025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.198384] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.198778] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22495f56-c154-496b-9163-396868977a1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.208296] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 933.208296] env[62558]: value = "task-1267039" [ 933.208296] env[62558]: _type = "Task" [ 933.208296] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.218957] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.249827] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.296776] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 933.329652] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.389095] env[62558]: DEBUG oslo_concurrency.lockutils [req-4643c814-d575-4cf7-aefd-5c8596c51884 req-10b78174-dd05-4bfe-8d5a-5f0795476245 service nova] Releasing lock "refresh_cache-a3f42034-3d2b-472a-89c0-5445cb6fb567" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.449014] env[62558]: DEBUG nova.network.neutron [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.455861] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.463707] env[62558]: DEBUG nova.compute.utils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.469157] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.469364] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.524163] env[62558]: DEBUG nova.policy [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.587083] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267037, 'name': ReconfigVM_Task, 'duration_secs': 1.152193} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.587523] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272562', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'name': 'volume-5dc54ac6-cb87-4624-a362-73c54c37b21e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2480822a-e10a-4066-a5d8-5ca633ab9b12', 'attached_at': '', 'detached_at': '', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'serial': '5dc54ac6-cb87-4624-a362-73c54c37b21e'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 933.588526] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eca0be8-edcf-423e-9bc2-e28baa0907ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.603492] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 933.603492] env[62558]: value = "task-1267040" [ 933.603492] env[62558]: _type = "Task" [ 933.603492] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.614445] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267040, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.722907] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267039, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.753719] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.833764] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.859208] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6a422c-1c97-459d-bd83-f825ef569093 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.868768] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9a25cc-d386-4cef-9168-5c6793a6e957 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.903783] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Successfully created port: 2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.906513] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8373eeba-aed2-47f5-bca8-904c7bae6609 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.916012] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a8c8aa-4936-495c-a4a2-ac431b6f4f93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.936214] env[62558]: DEBUG nova.compute.provider_tree [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.952226] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.952551] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance network_info: |[{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.954603] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:be:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01b4d7b4-efd1-4cc4-aa28-2663beeb66be', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.965501] env[62558]: DEBUG oslo.service.loopingcall [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.973346] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.973346] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.974158] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.977899] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9eb054b0-aa00-4d78-a8fb-60da907d7dae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.004265] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.004265] env[62558]: value = "task-1267041" [ 934.004265] env[62558]: _type = "Task" [ 934.004265] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.015450] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267041, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.114238] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267040, 'name': Rename_Task, 'duration_secs': 0.174813} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.114597] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.114902] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-167cb99f-3d1e-43dc-a862-e61859e664dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.122118] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 934.122118] env[62558]: value = "task-1267042" [ 934.122118] env[62558]: _type = "Task" [ 934.122118] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.130721] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.222673] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.223100] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.223492] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.224050] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b1a147a-8f96-44de-b8e2-f0cb38f2277f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.233758] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 934.233758] env[62558]: value = "task-1267043" [ 934.233758] env[62558]: _type = "Task" [ 934.233758] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.246258] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.258728] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.296704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.297638] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.335406] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.440400] env[62558]: DEBUG nova.scheduler.client.report [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.458959] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.518031] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267041, 'name': CreateVM_Task, 'duration_secs': 0.422539} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.518941] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.519818] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.520081] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.520509] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.521176] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f2f65d5-e69c-4df7-ad1b-8c4858e1e422 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.527056] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 934.527056] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527f2645-6b08-414c-2a54-e323b60ff12a" [ 934.527056] env[62558]: _type = "Task" [ 934.527056] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.536839] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527f2645-6b08-414c-2a54-e323b60ff12a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.633329] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267042, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.745610] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084605} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.746048] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.750448] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3cedec-7646-4b43-b6f5-377b269743f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.758996] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.775369] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.775772] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45eeeb61-4f7d-419b-91ec-72da5c189c21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.796548] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 934.796548] env[62558]: value = "task-1267044" [ 934.796548] env[62558]: _type = "Task" [ 934.796548] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.800739] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.811801] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267044, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.832482] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.968535] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.985786] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.017780] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.018126] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.018303] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.018495] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.018678] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.018912] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.019222] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.019409] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.019591] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.019791] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.019985] env[62558]: DEBUG nova.virt.hardware [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.021029] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e19af23-971b-487b-b19d-73a08000a1e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.034829] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab85584e-5a04-48a7-9b18-49274f135340 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.046370] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527f2645-6b08-414c-2a54-e323b60ff12a, 'name': SearchDatastore_Task, 'duration_secs': 0.01587} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.055155] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.055567] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.055972] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.056301] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.057018] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.057183] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3489ac25-e167-48ae-a9e0-92701f96db10 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.068511] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.068750] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.070402] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-899a0ae0-e1a1-437b-a006-89c07ee84690 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.077967] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 935.077967] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528035fd-423b-f011-cdb7-33c9a1e287c5" [ 935.077967] env[62558]: _type = "Task" [ 935.077967] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.089322] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528035fd-423b-f011-cdb7-33c9a1e287c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.134116] env[62558]: DEBUG oslo_vmware.api [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267042, 'name': PowerOnVM_Task, 'duration_secs': 0.616236} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.134502] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.134732] env[62558]: INFO nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Took 5.50 seconds to spawn the instance on the hypervisor. [ 935.134898] env[62558]: DEBUG nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.135754] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25148ed7-c09a-4fe9-b9fb-338c693990d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.257118] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267034, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.056052} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.257659] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a/OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a.vmdk to [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk. [ 935.257659] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Cleaning up location [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 935.257941] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_524e6a29-72dc-4e45-8c03-d5839e3ead4a {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.258122] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7a6394c-d1a3-4474-ac13-092e9fa2f227 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.264847] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 935.264847] env[62558]: value = "task-1267045" [ 935.264847] env[62558]: _type = "Task" [ 935.264847] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.274211] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267045, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.310031] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267044, 'name': ReconfigVM_Task, 'duration_secs': 0.487812} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.312631] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Reconfigured VM instance instance-00000057 to attach disk [datastore1] f8d6df86-52b8-4692-add4-1ffec82cc598/f8d6df86-52b8-4692-add4-1ffec82cc598.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.313630] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f353589-1e2c-4987-92ff-8d4654be05a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.320227] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 935.320227] env[62558]: value = "task-1267046" [ 935.320227] env[62558]: _type = "Task" [ 935.320227] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.333690] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267046, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.343936] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.344711] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.462720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.504s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.466644] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.467278] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.083s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.469750] env[62558]: INFO nova.compute.claims [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.532939] env[62558]: DEBUG nova.compute.manager [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Received event network-vif-plugged-2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.533281] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] Acquiring lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.533546] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.533733] env[62558]: DEBUG oslo_concurrency.lockutils [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.533932] env[62558]: DEBUG nova.compute.manager [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] No waiting events found dispatching network-vif-plugged-2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.534069] env[62558]: WARNING nova.compute.manager [req-a7fd2c27-da52-4430-bdc6-6180bcc83f21 req-acdad912-6224-49c3-a481-2229c36fa024 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Received unexpected event network-vif-plugged-2fc3029c-f3d0-4f21-91c0-1328caff5374 for instance with vm_state building and task_state spawning. [ 935.553909] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Successfully updated port: 2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.589907] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528035fd-423b-f011-cdb7-33c9a1e287c5, 'name': SearchDatastore_Task, 'duration_secs': 0.021572} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.590840] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79678898-2f88-40b2-a694-c7f5b06b81d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.596874] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 935.596874] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4419d-518b-431a-c575-a2f58cf3d23c" [ 935.596874] env[62558]: _type = "Task" [ 935.596874] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.607588] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4419d-518b-431a-c575-a2f58cf3d23c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.656028] env[62558]: INFO nova.compute.manager [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Took 27.08 seconds to build instance. [ 935.775992] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267045, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296443} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.776943] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.777235] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.777533] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk to [datastore2] 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6/42b2e70f-a46b-4984-b6d1-a2bfe348d1f6.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.777877] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d0400a9-eb8a-4948-ba34-b9734fc2d8c8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.785376] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 935.785376] env[62558]: value = "task-1267047" [ 935.785376] env[62558]: _type = "Task" [ 935.785376] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.796211] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.834372] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267046, 'name': Rename_Task, 'duration_secs': 0.208264} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.838562] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.838938] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.839239] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-555e09dd-81d2-45c0-8101-88b37cb3ecdb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.845855] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 935.845855] env[62558]: value = "task-1267048" [ 935.845855] env[62558]: _type = "Task" [ 935.845855] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.854861] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267048, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.961987] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.034470] env[62558]: INFO nova.scheduler.client.report [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocation for migration d5cc95be-a6ee-4246-bcb7-88ff1cd2f615 [ 936.059050] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.059287] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.059287] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.108475] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52b4419d-518b-431a-c575-a2f58cf3d23c, 'name': SearchDatastore_Task, 'duration_secs': 0.017127} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.109311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.109311] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 29584082-37e6-4dd8-906f-fd87b4ca9bc2/29584082-37e6-4dd8-906f-fd87b4ca9bc2.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 936.109603] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9c4c607-f1e4-4a67-9024-48628bc03b9b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.117688] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 936.117688] env[62558]: value = "task-1267049" [ 936.117688] env[62558]: _type = "Task" [ 936.117688] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.131465] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.158396] env[62558]: DEBUG oslo_concurrency.lockutils [None req-50ad35b3-fe05-4158-9fd8-edfb5b1369ca tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.592s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.302015] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.340031] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267038, 'name': ReconfigVM_Task, 'duration_secs': 3.171525} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.340031] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de/bd09177c-5fd7-495c-893b-5ec219b6b3de-rescue.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.340031] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fcd302-26c5-4eb4-ae7c-53d250d4f44b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.372795] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f19b1fea-2104-4c50-97c2-b8496c88d24b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.389884] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267048, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.391754] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 936.391754] env[62558]: value = "task-1267050" [ 936.391754] env[62558]: _type = "Task" [ 936.391754] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.401011] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.466221] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267036, 'name': CopyVirtualDisk_Task, 'duration_secs': 4.336972} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.466597] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2f0f0be9-1e05-4998-a402-10590e152a59/2f0f0be9-1e05-4998-a402-10590e152a59.vmdk to [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.467540] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504e6b31-36fc-4ac8-9876-8ddca40fc777 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.496086] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.496981] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6746b85a-bc66-467b-9c9d-5aeefb0359e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.520075] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 936.520075] env[62558]: value = "task-1267051" [ 936.520075] env[62558]: _type = "Task" [ 936.520075] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.533443] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267051, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.544862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-47f76ff1-9537-4182-995c-e9cd896601be tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 23.354s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.611045] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.627832] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267049, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.790411] env[62558]: DEBUG nova.network.neutron [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Updating instance_info_cache with network_info: [{"id": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "address": "fa:16:3e:37:cd:ba", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc3029c-f3", "ovs_interfaceid": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.811323] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.874748] env[62558]: DEBUG oslo_vmware.api [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267048, 'name': PowerOnVM_Task, 'duration_secs': 0.644507} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.876174] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.876453] env[62558]: DEBUG nova.compute.manager [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.877280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f87cc0-4834-4bbb-ba2f-a40dafe80002 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.880839] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e859812-ae4a-449d-ac95-cf6376371782 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.890129] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325d0d7a-9cca-4331-9d1d-c44d27667560 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.908239] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.937760] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64054edc-f96b-44d5-beed-88f800c3c8b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.948047] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c215e2-06bc-4c09-8ba5-b1ba94d00390 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.965904] env[62558]: DEBUG nova.compute.provider_tree [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.034409] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267051, 'name': ReconfigVM_Task, 'duration_secs': 0.496313} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.034817] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5/8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.035650] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c616e8b6-4cb5-49b5-b977-dc0f7efddc98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.043390] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 937.043390] env[62558]: value = "task-1267052" [ 937.043390] env[62558]: _type = "Task" [ 937.043390] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.058194] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267052, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.061445] env[62558]: DEBUG nova.compute.manager [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Received event network-changed-1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.061694] env[62558]: DEBUG nova.compute.manager [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Refreshing instance network info cache due to event network-changed-1f94c71a-6387-41af-be7e-7fb674e8b3af. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.061935] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] Acquiring lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.062188] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] Acquired lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.062400] env[62558]: DEBUG nova.network.neutron [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Refreshing network info cache for port 1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.129827] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517912} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.130652] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 29584082-37e6-4dd8-906f-fd87b4ca9bc2/29584082-37e6-4dd8-906f-fd87b4ca9bc2.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.130907] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.131237] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5db91b51-4c1b-488f-ae14-0f1b13a702f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.140016] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 937.140016] env[62558]: value = "task-1267053" [ 937.140016] env[62558]: _type = "Task" [ 937.140016] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.149518] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.302246] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.306306] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Instance network_info: |[{"id": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "address": "fa:16:3e:37:cd:ba", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc3029c-f3", "ovs_interfaceid": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.306306] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.306306] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:cd:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fc3029c-f3d0-4f21-91c0-1328caff5374', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.313164] env[62558]: DEBUG oslo.service.loopingcall [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.313510] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.313703] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1e8c21e-7908-45fe-b7ce-e21872c235ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.338931] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.338931] env[62558]: value = "task-1267054" [ 937.338931] env[62558]: _type = "Task" [ 937.338931] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.347957] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267054, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.414290] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.417196] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.469496] env[62558]: DEBUG nova.scheduler.client.report [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.554126] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267052, 'name': Rename_Task, 'duration_secs': 0.221443} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.554467] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.554745] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-603054a9-3484-425b-a019-edf227ff0a99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.561114] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 937.561114] env[62558]: value = "task-1267055" [ 937.561114] env[62558]: _type = "Task" [ 937.561114] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.570697] env[62558]: DEBUG nova.compute.manager [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Received event network-changed-2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.570859] env[62558]: DEBUG nova.compute.manager [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Refreshing instance network info cache due to event network-changed-2fc3029c-f3d0-4f21-91c0-1328caff5374. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.571012] env[62558]: DEBUG oslo_concurrency.lockutils [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] Acquiring lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.571169] env[62558]: DEBUG oslo_concurrency.lockutils [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] Acquired lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.571425] env[62558]: DEBUG nova.network.neutron [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Refreshing network info cache for port 2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.580119] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267055, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.652633] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.339567} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.652633] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.653031] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b967f7c-25d5-4fbc-a1c6-0055d93000fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.678183] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 29584082-37e6-4dd8-906f-fd87b4ca9bc2/29584082-37e6-4dd8-906f-fd87b4ca9bc2.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.681275] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfe94592-a537-4a78-9bfd-12d9aa339833 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.702657] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 937.702657] env[62558]: value = "task-1267056" [ 937.702657] env[62558]: _type = "Task" [ 937.702657] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.712556] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.802765] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.851588] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267054, 'name': CreateVM_Task, 'duration_secs': 0.420963} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.851869] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 937.852534] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.852714] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.853129] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 937.853375] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aece9e1-7eca-4cbf-8dde-bfc9406ef1fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.858648] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 937.858648] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5258ed7b-74bc-fc0e-2075-84eac9c42f80" [ 937.858648] env[62558]: _type = "Task" [ 937.858648] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.867697] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5258ed7b-74bc-fc0e-2075-84eac9c42f80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.902296] env[62558]: DEBUG nova.network.neutron [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updated VIF entry in instance network info cache for port 1f94c71a-6387-41af-be7e-7fb674e8b3af. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 937.902296] env[62558]: DEBUG nova.network.neutron [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating instance_info_cache with network_info: [{"id": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "address": "fa:16:3e:08:60:a5", "network": {"id": "adf89e17-7cea-47a6-8654-a97be0414a6c", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1824548634-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d567aaa721d04cac8a80a34c7104fb05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f94c71a-63", "ovs_interfaceid": "1f94c71a-6387-41af-be7e-7fb674e8b3af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.912103] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267050, 'name': ReconfigVM_Task, 'duration_secs': 1.146761} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.912912] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.913419] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e50864d1-65da-49cc-96b4-db4cf2b04709 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.921831] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 937.921831] env[62558]: value = "task-1267057" [ 937.921831] env[62558]: _type = "Task" [ 937.921831] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.932464] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.977921] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.977921] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.981204] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.483s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.982351] env[62558]: DEBUG nova.objects.instance [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'resources' on Instance uuid ecf01ab6-5019-4453-b102-0e754abc4ef8 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.076772] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267055, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.219446] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267056, 'name': ReconfigVM_Task, 'duration_secs': 0.491621} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.220149] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 29584082-37e6-4dd8-906f-fd87b4ca9bc2/29584082-37e6-4dd8-906f-fd87b4ca9bc2.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.221619] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2a35e30-3681-43f5-bf06-819a2b790863 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.231027] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 938.231027] env[62558]: value = "task-1267058" [ 938.231027] env[62558]: _type = "Task" [ 938.231027] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.244405] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267058, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.307927] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.340087] env[62558]: DEBUG nova.network.neutron [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Updated VIF entry in instance network info cache for port 2fc3029c-f3d0-4f21-91c0-1328caff5374. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.340537] env[62558]: DEBUG nova.network.neutron [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Updating instance_info_cache with network_info: [{"id": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "address": "fa:16:3e:37:cd:ba", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc3029c-f3", "ovs_interfaceid": "2fc3029c-f3d0-4f21-91c0-1328caff5374", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.370974] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5258ed7b-74bc-fc0e-2075-84eac9c42f80, 'name': SearchDatastore_Task, 'duration_secs': 0.014851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.371349] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.371611] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.371889] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.372133] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.372273] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.372585] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b702bb1-8954-4392-b1ff-5dab94582bef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.386054] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.386332] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.387616] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70a8a6ee-b292-4e6a-ab69-33707aed3fc8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.393983] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 938.393983] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5228e9e1-1fc2-9736-f23f-a998bb749282" [ 938.393983] env[62558]: _type = "Task" [ 938.393983] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.405043] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5228e9e1-1fc2-9736-f23f-a998bb749282, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.406790] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1402d03-c11b-4b5c-bd39-18ba4ccb6290 req-4a74ddbb-39a2-47e6-8ec1-3380ea7fd603 service nova] Releasing lock "refresh_cache-2480822a-e10a-4066-a5d8-5ca633ab9b12" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.432362] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267057, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.472826] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 938.473904] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da02446-d204-4490-aa4e-3fae29ed693a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.480839] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 938.481041] env[62558]: ERROR oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk due to incomplete transfer. [ 938.481328] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b4850f9f-9d4e-47c0-881e-5f29f12c760f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.485465] env[62558]: DEBUG nova.compute.utils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.487150] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.487321] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.493750] env[62558]: DEBUG oslo_vmware.rw_handles [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52210f37-4549-80dc-7381-5db3926e2183/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 938.493971] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Uploaded image bf57715a-d9cb-4e6c-885f-8cbb16b14efd to the Glance image server {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 938.496417] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 938.497027] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4b61a98e-346f-4e58-b0b6-04a542d4e668 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.503675] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 938.503675] env[62558]: value = "task-1267059" [ 938.503675] env[62558]: _type = "Task" [ 938.503675] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.515565] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267059, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.537801] env[62558]: DEBUG nova.policy [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6780d34174af44fc834709fff7dbc64a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f8a5c4ed9e04fc69280ab0b034c8c3d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.571856] env[62558]: DEBUG oslo_vmware.api [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267055, 'name': PowerOnVM_Task, 'duration_secs': 0.700083} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.571856] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.709021] env[62558]: DEBUG nova.compute.manager [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.713016] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af861fcd-4422-4779-a0d3-0020cad1b920 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.745441] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267058, 'name': Rename_Task, 'duration_secs': 0.186962} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.748953] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.749545] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b28ad79-4ea8-46fe-8701-44a1065684be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.757742] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 938.757742] env[62558]: value = "task-1267060" [ 938.757742] env[62558]: _type = "Task" [ 938.757742] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.772863] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.780422] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f8d6df86-52b8-4692-add4-1ffec82cc598" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.780668] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.780874] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f8d6df86-52b8-4692-add4-1ffec82cc598-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.781080] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.781246] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.783473] env[62558]: INFO nova.compute.manager [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Terminating instance [ 938.785519] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "refresh_cache-f8d6df86-52b8-4692-add4-1ffec82cc598" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.785695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "refresh_cache-f8d6df86-52b8-4692-add4-1ffec82cc598" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.786147] env[62558]: DEBUG nova.network.neutron [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.808420] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267047, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.86854} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.808981] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe8845c5-445d-4f71-8ce6-a1a0ce770a67/fe8845c5-445d-4f71-8ce6-a1a0ce770a67.vmdk to [datastore2] 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6/42b2e70f-a46b-4984-b6d1-a2bfe348d1f6.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.809917] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84201746-5888-4dac-817a-fc32b59f7f2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.834169] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6/42b2e70f-a46b-4984-b6d1-a2bfe348d1f6.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.837596] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Successfully created port: 253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.839405] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2d0cb4f-ce42-4e9f-a928-6d2516801881 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.856236] env[62558]: DEBUG oslo_concurrency.lockutils [req-8f36ec1c-239d-416c-9d46-a1b3aa720b1b req-a954a4ea-045a-4383-8a31-8dce907fa781 service nova] Releasing lock "refresh_cache-ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.863215] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 938.863215] env[62558]: value = "task-1267061" [ 938.863215] env[62558]: _type = "Task" [ 938.863215] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.872954] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267061, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.911715] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5228e9e1-1fc2-9736-f23f-a998bb749282, 'name': SearchDatastore_Task, 'duration_secs': 0.01422} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.915257] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-599499d6-ed28-4d66-a4ec-1ce796eaf804 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.924114] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 938.924114] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff2d95-ecd3-7832-5158-5cf27c07a3fb" [ 938.924114] env[62558]: _type = "Task" [ 938.924114] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.934523] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff2d95-ecd3-7832-5158-5cf27c07a3fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.940908] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267057, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.993412] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.997915] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cc2c6c-e173-455f-ba76-bf54f0a29da8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.013555] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b678df-246b-41a9-b871-db58e22b088b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.021899] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267059, 'name': Destroy_Task, 'duration_secs': 0.316338} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.022600] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Destroyed the VM [ 939.022872] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 939.023314] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ba5b860-d30b-4b49-8e0b-f14b455ef094 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.055675] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c84d3fc-8c98-4ec0-b1d6-ec928cc25220 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.060541] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 939.060541] env[62558]: value = "task-1267062" [ 939.060541] env[62558]: _type = "Task" [ 939.060541] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.069040] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb3efc1-6b3a-4484-9fd0-31587232d6b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.077862] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267062, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.088449] env[62558]: DEBUG nova.compute.provider_tree [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.229071] env[62558]: DEBUG oslo_concurrency.lockutils [None req-04db6982-b782-4dec-90f1-03364b2dd49c tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 36.704s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.267454] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267060, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.303452] env[62558]: DEBUG nova.network.neutron [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.354116] env[62558]: DEBUG nova.network.neutron [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.373705] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267061, 'name': ReconfigVM_Task, 'duration_secs': 0.328984} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.374080] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6/42b2e70f-a46b-4984-b6d1-a2bfe348d1f6.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.374745] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0378522c-29cb-4fcf-8fab-47328608a4a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.381679] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 939.381679] env[62558]: value = "task-1267063" [ 939.381679] env[62558]: _type = "Task" [ 939.381679] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.390947] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267063, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.436300] env[62558]: DEBUG oslo_vmware.api [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267057, 'name': PowerOnVM_Task, 'duration_secs': 1.016965} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.440437] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.442695] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff2d95-ecd3-7832-5158-5cf27c07a3fb, 'name': SearchDatastore_Task, 'duration_secs': 0.013408} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.443805] env[62558]: DEBUG nova.compute.manager [None req-512c4cfd-eb35-4948-a9f7-c073ff5c7772 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.444173] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.444455] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568/ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.445239] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84fc85a-eec4-4166-bed3-216014cecb9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.447949] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6dc7e0da-9663-44d3-9a55-5b0e35e498e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.457334] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 939.457334] env[62558]: value = "task-1267064" [ 939.457334] env[62558]: _type = "Task" [ 939.457334] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.465420] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267064, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.506875] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.507232] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.507496] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "63050323-047e-4d73-91ae-859467b4b5a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.507704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.508392] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.510796] env[62558]: INFO nova.compute.manager [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Terminating instance [ 939.515784] env[62558]: DEBUG nova.compute.manager [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.515995] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.516894] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74e3458-88ca-4a1f-8d4a-e8dd04644640 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.525153] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.525430] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-487e6837-6c20-406b-86b0-26adb71daf51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.534353] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 939.534353] env[62558]: value = "task-1267065" [ 939.534353] env[62558]: _type = "Task" [ 939.534353] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.542135] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267065, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.571487] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267062, 'name': RemoveSnapshot_Task, 'duration_secs': 0.349404} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.571905] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 939.572141] env[62558]: DEBUG nova.compute.manager [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.572873] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c24ca2-3d3b-4a78-8412-42e5e70fed47 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.592511] env[62558]: DEBUG nova.scheduler.client.report [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.769948] env[62558]: DEBUG oslo_vmware.api [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267060, 'name': PowerOnVM_Task, 'duration_secs': 0.988264} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.769948] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.769948] env[62558]: INFO nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Took 9.38 seconds to spawn the instance on the hypervisor. [ 939.769948] env[62558]: DEBUG nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.769948] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ae7d5d-ab9c-4278-9f11-e490b2c46948 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.856657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "refresh_cache-f8d6df86-52b8-4692-add4-1ffec82cc598" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.857135] env[62558]: DEBUG nova.compute.manager [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.857337] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.858484] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f64047e-cf30-4c45-92e9-9ea402a46f4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.866077] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.866274] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed5aadcb-7aca-4215-be68-86d93dfabc01 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.872643] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 939.872643] env[62558]: value = "task-1267066" [ 939.872643] env[62558]: _type = "Task" [ 939.872643] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.880552] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267066, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.891372] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267063, 'name': Rename_Task, 'duration_secs': 0.155347} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.891655] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.891894] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b247c544-ecd3-441a-b935-d32f9867e7c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.899077] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 939.899077] env[62558]: value = "task-1267067" [ 939.899077] env[62558]: _type = "Task" [ 939.899077] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.907604] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.973419] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267064, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.011908] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.044276] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.044709] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.044953] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.045265] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.045525] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.045771] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.046412] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.046784] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.046930] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.047183] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.047778] env[62558]: DEBUG nova.virt.hardware [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.049631] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e986cfd4-5b38-4118-aa64-67ebfcde1dfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.060547] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267065, 'name': PowerOffVM_Task, 'duration_secs': 0.189503} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.065705] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.066178] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.070506] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96e1efae-8986-4adf-accb-907b931beb2c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.074634] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a832fb-8835-40d0-96de-d8b4862d9931 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.083331] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.083624] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.089947] env[62558]: INFO nova.compute.manager [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Shelve offloading [ 940.105819] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.110658] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.111443] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.019s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.111774] env[62558]: DEBUG nova.objects.instance [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lazy-loading 'resources' on Instance uuid 7aebdb8a-517b-4168-91e0-1b704e6a11c7 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.113238] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7345d7a-146c-44fd-9ef4-5342781a2ccd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.122012] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 940.122012] env[62558]: value = "task-1267069" [ 940.122012] env[62558]: _type = "Task" [ 940.122012] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.136511] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 940.136912] env[62558]: DEBUG nova.compute.manager [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.138280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ae7ab8-11b7-45ad-97d7-2e94becc33c8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.143188] env[62558]: INFO nova.scheduler.client.report [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance ecf01ab6-5019-4453-b102-0e754abc4ef8 [ 940.153185] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.153185] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.153185] env[62558]: DEBUG nova.network.neutron [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.155783] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.156074] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.156360] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleting the datastore file [datastore2] 63050323-047e-4d73-91ae-859467b4b5a7 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.156966] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc02769f-beb9-468d-a700-89179e1508dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.165634] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 940.165634] env[62558]: value = "task-1267070" [ 940.165634] env[62558]: _type = "Task" [ 940.165634] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.176873] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.287655] env[62558]: INFO nova.compute.manager [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Took 29.43 seconds to build instance. [ 940.316636] env[62558]: INFO nova.compute.manager [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Unrescuing [ 940.316879] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.317572] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.317572] env[62558]: DEBUG nova.network.neutron [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.333590] env[62558]: DEBUG nova.compute.manager [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Received event network-vif-plugged-253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.333806] env[62558]: DEBUG oslo_concurrency.lockutils [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] Acquiring lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.334023] env[62558]: DEBUG oslo_concurrency.lockutils [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.334228] env[62558]: DEBUG oslo_concurrency.lockutils [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.334407] env[62558]: DEBUG nova.compute.manager [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] No waiting events found dispatching network-vif-plugged-253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.334573] env[62558]: WARNING nova.compute.manager [req-cf4bd52f-d645-4746-b3d1-2295bb4bd7b9 req-8c910c85-6d99-421c-a9c1-52df3949c3e5 service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Received unexpected event network-vif-plugged-253ac594-cda6-41b3-a77e-c8ce05524fbd for instance with vm_state building and task_state spawning. [ 940.383896] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267066, 'name': PowerOffVM_Task, 'duration_secs': 0.224944} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.384246] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.384523] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.385580] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfbbe239-44b2-4e82-874d-d202e58e4f67 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.410614] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267067, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.411860] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.412074] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.412260] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleting the datastore file [datastore1] f8d6df86-52b8-4692-add4-1ffec82cc598 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.412513] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a057e92-0edc-451e-bc31-a194507d22ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.419523] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 940.419523] env[62558]: value = "task-1267072" [ 940.419523] env[62558]: _type = "Task" [ 940.419523] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.427864] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.434545] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Successfully updated port: 253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.471396] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267064, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.767298} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.471593] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568/ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.471804] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.472262] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-edcb3094-330f-41eb-80e0-3f0fcb089e8d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.479343] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 940.479343] env[62558]: value = "task-1267073" [ 940.479343] env[62558]: _type = "Task" [ 940.479343] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.487792] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267073, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.589703] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.660207] env[62558]: DEBUG oslo_concurrency.lockutils [None req-99bc5f25-b826-4117-85fd-4c64b0999698 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "ecf01ab6-5019-4453-b102-0e754abc4ef8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.529s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.675756] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267070, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.789558] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0115c8ac-7b4c-441f-88f4-4e5b685d95bf tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.945s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.882849] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5877246b-d2a9-4511-a956-ddab101cabc4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.891079] env[62558]: DEBUG nova.network.neutron [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.893220] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eca9b4-cf38-46b3-9f04-474f9099bed5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.933765] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb765c6-6390-48c0-a3a0-b0d146b8f49f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.940742] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.941152] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquired lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.941152] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.942992] env[62558]: DEBUG oslo_vmware.api [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267067, 'name': PowerOnVM_Task, 'duration_secs': 0.531809} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.946218] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.946218] env[62558]: INFO nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Took 16.52 seconds to spawn the instance on the hypervisor. [ 940.947363] env[62558]: DEBUG nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.952140] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74944372-9181-45af-b383-2d4d825e3642 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.956141] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.960157] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894e70af-2cbe-4c09-90e3-8661256daf03 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.981714] env[62558]: DEBUG nova.compute.provider_tree [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.993265] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267073, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068192} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.994119] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.994959] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffcf9f9-2fea-4e94-b8aa-acc51de8ca11 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.019007] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568/ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.022035] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-565a7d82-b626-422d-acd3-9834fbb6fe7e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.042725] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 941.042725] env[62558]: value = "task-1267074" [ 941.042725] env[62558]: _type = "Task" [ 941.042725] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.055398] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267074, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.119553] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.150635] env[62558]: DEBUG nova.network.neutron [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.180536] env[62558]: DEBUG oslo_vmware.api [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267070, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.73031} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.180799] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.181726] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.181726] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.181726] env[62558]: INFO nova.compute.manager [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Took 1.67 seconds to destroy the instance on the hypervisor. [ 941.181726] env[62558]: DEBUG oslo.service.loopingcall [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.181951] env[62558]: DEBUG nova.compute.manager [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.182438] env[62558]: DEBUG nova.network.neutron [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.400403] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.441109] env[62558]: DEBUG oslo_vmware.api [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.536132} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.441109] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.441556] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 941.441932] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 941.442954] env[62558]: INFO nova.compute.manager [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Took 1.58 seconds to destroy the instance on the hypervisor. [ 941.442954] env[62558]: DEBUG oslo.service.loopingcall [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.443353] env[62558]: DEBUG nova.compute.manager [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.443584] env[62558]: DEBUG nova.network.neutron [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 941.480283] env[62558]: INFO nova.compute.manager [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Took 34.23 seconds to build instance. [ 941.488359] env[62558]: DEBUG nova.network.neutron [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.490024] env[62558]: DEBUG nova.scheduler.client.report [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.518742] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.563196] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267074, 'name': ReconfigVM_Task, 'duration_secs': 0.459889} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.563521] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Reconfigured VM instance instance-0000005b to attach disk [datastore1] ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568/ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.564410] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c812aad-9cad-4860-8292-704fbc231f90 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.570755] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 941.570755] env[62558]: value = "task-1267075" [ 941.570755] env[62558]: _type = "Task" [ 941.570755] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.582612] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267075, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.656893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.657975] env[62558]: DEBUG nova.objects.instance [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'flavor' on Instance uuid 4815ba3f-265f-466a-9850-4c325cdb88de {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.683605] env[62558]: DEBUG nova.compute.manager [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-changed-01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.683726] env[62558]: DEBUG nova.compute.manager [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing instance network info cache due to event network-changed-01b4d7b4-efd1-4cc4-aa28-2663beeb66be. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 941.683921] env[62558]: DEBUG oslo_concurrency.lockutils [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.684522] env[62558]: DEBUG oslo_concurrency.lockutils [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.684874] env[62558]: DEBUG nova.network.neutron [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing network info cache for port 01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.743435] env[62558]: DEBUG nova.network.neutron [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Updating instance_info_cache with network_info: [{"id": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "address": "fa:16:3e:1b:0c:34", "network": {"id": "c80ac726-7352-4f2f-9667-01d964af7e85", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-545406572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f8a5c4ed9e04fc69280ab0b034c8c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f41e4aa-0d23-48c4-a359-574abb2e7b9a", "external-id": "nsx-vlan-transportzone-695", "segmentation_id": 695, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap253ac594-cd", "ovs_interfaceid": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.759828] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.760758] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f653147-6bb3-437e-a028-05e749028864 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.774067] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.774314] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d947683c-6746-416a-8860-322511379ca8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.860941] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.861205] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.861392] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleting the datastore file [datastore2] 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.861671] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ec9379e-f43f-4c1f-a1d5-ea43e59ac1e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.870024] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 941.870024] env[62558]: value = "task-1267077" [ 941.870024] env[62558]: _type = "Task" [ 941.870024] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.878450] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.982650] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1004bb6d-ad0b-4d93-b456-7783ff271c5a tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.757s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.996150] env[62558]: DEBUG nova.network.neutron [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.997920] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.007620] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.663s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.009348] env[62558]: INFO nova.compute.claims [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.020248] env[62558]: INFO nova.scheduler.client.report [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Deleted allocations for instance 7aebdb8a-517b-4168-91e0-1b704e6a11c7 [ 942.059406] env[62558]: DEBUG nova.network.neutron [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.083911] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267075, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.163780] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea168253-4579-4841-abe6-03501197a3ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.187017] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.187362] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad5b67a7-819c-47c5-9bd7-0a398587dbc0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.195667] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 942.195667] env[62558]: value = "task-1267078" [ 942.195667] env[62558]: _type = "Task" [ 942.195667] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.204304] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.246601] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Releasing lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.247034] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Instance network_info: |[{"id": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "address": "fa:16:3e:1b:0c:34", "network": {"id": "c80ac726-7352-4f2f-9667-01d964af7e85", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-545406572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f8a5c4ed9e04fc69280ab0b034c8c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f41e4aa-0d23-48c4-a359-574abb2e7b9a", "external-id": "nsx-vlan-transportzone-695", "segmentation_id": 695, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap253ac594-cd", "ovs_interfaceid": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.247513] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:0c:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f41e4aa-0d23-48c4-a359-574abb2e7b9a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '253ac594-cda6-41b3-a77e-c8ce05524fbd', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.255640] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Creating folder: Project (0f8a5c4ed9e04fc69280ab0b034c8c3d). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.256118] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67f8eccb-1b86-47b1-8478-f05a5da2cb40 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.269405] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Created folder: Project (0f8a5c4ed9e04fc69280ab0b034c8c3d) in parent group-v272451. [ 942.269613] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Creating folder: Instances. Parent ref: group-v272588. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.271103] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.271334] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.271537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.271719] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.271918] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.273478] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24f436cb-1b5b-4b13-a499-a29cc42465d5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.277853] env[62558]: INFO nova.compute.manager [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Terminating instance [ 942.279419] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "591d326c-c20c-401b-a57b-895d59c3b418" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.279594] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.280771] env[62558]: DEBUG nova.compute.manager [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.280954] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.282076] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7bea26-e6d4-4494-9dc5-5fa8e1edecff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.288879] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Created folder: Instances in parent group-v272588. [ 942.289034] env[62558]: DEBUG oslo.service.loopingcall [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.289523] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.289756] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccd4a8d6-7c0c-4f8a-b965-e4480cfc0911 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.308840] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.309754] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a0fb5d2-1af7-43ca-b1bb-c2a5d13e1105 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.313637] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.313637] env[62558]: value = "task-1267081" [ 942.313637] env[62558]: _type = "Task" [ 942.313637] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.318197] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 942.318197] env[62558]: value = "task-1267082" [ 942.318197] env[62558]: _type = "Task" [ 942.318197] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.323852] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267081, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.328609] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.368837] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Received event network-changed-253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.369466] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Refreshing instance network info cache due to event network-changed-253ac594-cda6-41b3-a77e-c8ce05524fbd. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.369466] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Acquiring lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.369466] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Acquired lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.369761] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Refreshing network info cache for port 253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.381746] env[62558]: DEBUG oslo_vmware.api [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.391708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.381746] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.381961] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.382134] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.399280] env[62558]: INFO nova.scheduler.client.report [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted allocations for instance 49a58b46-207f-4515-b313-afcdb2d1ced3 [ 942.447850] env[62558]: DEBUG nova.network.neutron [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updated VIF entry in instance network info cache for port 01b4d7b4-efd1-4cc4-aa28-2663beeb66be. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.448400] env[62558]: DEBUG nova.network.neutron [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.508410] env[62558]: INFO nova.compute.manager [-] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Took 1.06 seconds to deallocate network for instance. [ 942.529797] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0b8690fc-8ee5-4e7f-b18b-c0b191fe32f0 tempest-ServerRescueTestJSONUnderV235-2029275303 tempest-ServerRescueTestJSONUnderV235-2029275303-project-member] Lock "7aebdb8a-517b-4168-91e0-1b704e6a11c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.585s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.561705] env[62558]: INFO nova.compute.manager [-] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Took 1.38 seconds to deallocate network for instance. [ 942.581780] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267075, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.707046] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267078, 'name': PowerOffVM_Task, 'duration_secs': 0.198155} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.707046] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.712215] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfiguring VM instance instance-0000004e to detach disk 2002 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 942.712510] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a340c6b-8e5f-43fd-9728-475e8614d904 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.732410] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 942.732410] env[62558]: value = "task-1267083" [ 942.732410] env[62558]: _type = "Task" [ 942.732410] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.741676] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.789534] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.826418] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267081, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.832769] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267082, 'name': PowerOffVM_Task, 'duration_secs': 0.387444} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.833283] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.833572] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.833940] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49091a93-4321-495b-b2b8-25fe89882694 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.907137] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.942743] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.943087] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.943421] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleting the datastore file [datastore2] 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.943815] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a729f90-a5a6-4743-9cbb-92feaa63e6dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.951081] env[62558]: DEBUG oslo_concurrency.lockutils [req-6513fe68-ae88-4dbe-b4b9-673290e10581 req-16d13f30-72dd-4a29-8969-0cc66c2524f3 service nova] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.953409] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 942.953409] env[62558]: value = "task-1267085" [ 942.953409] env[62558]: _type = "Task" [ 942.953409] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.965947] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.017158] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.069640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.086137] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267075, 'name': Rename_Task, 'duration_secs': 1.110991} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.090788] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.091326] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95cafbd7-c882-47e5-81c0-7f38e1876652 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.098249] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 943.098249] env[62558]: value = "task-1267086" [ 943.098249] env[62558]: _type = "Task" [ 943.098249] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.109318] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.127542] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Updated VIF entry in instance network info cache for port 253ac594-cda6-41b3-a77e-c8ce05524fbd. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.128101] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Updating instance_info_cache with network_info: [{"id": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "address": "fa:16:3e:1b:0c:34", "network": {"id": "c80ac726-7352-4f2f-9667-01d964af7e85", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-545406572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f8a5c4ed9e04fc69280ab0b034c8c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f41e4aa-0d23-48c4-a359-574abb2e7b9a", "external-id": "nsx-vlan-transportzone-695", "segmentation_id": 695, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap253ac594-cd", "ovs_interfaceid": "253ac594-cda6-41b3-a77e-c8ce05524fbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.246022] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267083, 'name': ReconfigVM_Task, 'duration_secs': 0.246689} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.246022] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfigured VM instance instance-0000004e to detach disk 2002 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.246022] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.246022] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1d80f1f-4b46-481f-9a93-b110d574e048 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.254892] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 943.254892] env[62558]: value = "task-1267087" [ 943.254892] env[62558]: _type = "Task" [ 943.254892] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.263423] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.299186] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81958174-fb9d-41fa-a9db-ef4c0507524d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.311716] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9be1b24-5bc0-4474-879e-b41db26e9888 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.317433] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.327664] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267081, 'name': CreateVM_Task, 'duration_secs': 0.601759} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.351268] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.352288] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.352465] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.352852] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.353669] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b438a9b8-b02a-42ec-a88d-1277d66a8b1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.356301] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99da8096-a008-4e58-af42-791f7118ff7a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.364031] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f6d074-dc12-45e8-9da2-03febb0c1789 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.368011] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 943.368011] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521eebcf-af76-c6f4-4a1d-55a5593c00af" [ 943.368011] env[62558]: _type = "Task" [ 943.368011] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.380723] env[62558]: DEBUG nova.compute.provider_tree [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.387169] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521eebcf-af76-c6f4-4a1d-55a5593c00af, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.387730] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.387998] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.388307] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.388603] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.388683] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.389315] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1087e15-8ef6-47fb-b445-4e818e1a6213 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.396929] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.397123] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.397832] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5926b379-2cbd-4799-96a7-fc8f123687e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.404034] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 943.404034] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52573f12-07be-c48b-defb-5eff77dd00bb" [ 943.404034] env[62558]: _type = "Task" [ 943.404034] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.412029] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52573f12-07be-c48b-defb-5eff77dd00bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.464423] env[62558]: DEBUG oslo_vmware.api [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173288} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.464746] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.464973] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.465200] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.465420] env[62558]: INFO nova.compute.manager [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 943.465688] env[62558]: DEBUG oslo.service.loopingcall [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.465881] env[62558]: DEBUG nova.compute.manager [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 943.465976] env[62558]: DEBUG nova.network.neutron [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 943.608439] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267086, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.633446] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Releasing lock "refresh_cache-7188fa81-3a23-4f22-adb8-41a35b1af9a5" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.633835] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Received event network-vif-deleted-174e00d0-f0e5-4b0a-9ca2-4db08bb7f674 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.634784] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-vif-unplugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.634784] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.634784] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.634784] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.634784] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] No waiting events found dispatching network-vif-unplugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.634965] env[62558]: WARNING nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received unexpected event network-vif-unplugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a for instance with vm_state shelved and task_state shelving_offloading. [ 943.635114] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.635444] env[62558]: DEBUG nova.compute.manager [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing instance network info cache due to event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.635544] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.635638] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.635799] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.765798] env[62558]: DEBUG oslo_vmware.api [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267087, 'name': PowerOnVM_Task, 'duration_secs': 0.39207} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.766216] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.766525] env[62558]: DEBUG nova.compute.manager [None req-0e45f6fc-7b64-4294-88c4-c12d84f6e255 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.767428] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385af688-b939-4d8a-9f9d-0c9e1f1a1c56 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.888691] env[62558]: DEBUG nova.scheduler.client.report [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.915349] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52573f12-07be-c48b-defb-5eff77dd00bb, 'name': SearchDatastore_Task, 'duration_secs': 0.008042} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.916156] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b90771a-0cae-42dc-86f7-1735958ab3f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.921441] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 943.921441] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e6a79e-8e10-9280-ed5d-06c2911f2fda" [ 943.921441] env[62558]: _type = "Task" [ 943.921441] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.929222] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e6a79e-8e10-9280-ed5d-06c2911f2fda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.110708] env[62558]: DEBUG oslo_vmware.api [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267086, 'name': PowerOnVM_Task, 'duration_secs': 0.774287} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.111156] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.111409] env[62558]: INFO nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Took 9.13 seconds to spawn the instance on the hypervisor. [ 944.111651] env[62558]: DEBUG nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.112514] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6d99a6-0021-4fef-801a-546491c23b40 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.188708] env[62558]: DEBUG nova.network.neutron [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.393281] env[62558]: DEBUG nova.compute.manager [req-90fefb09-a604-45e4-8fa3-8a6233f05dca req-65dba15e-cfcd-4f7f-8679-9a2f342797ad service nova] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Received event network-vif-deleted-d08a37d6-d27f-467e-9138-f2ff9de7f070 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.394547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.395223] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.399703] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.983s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.404716] env[62558]: DEBUG nova.objects.instance [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 944.436431] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e6a79e-8e10-9280-ed5d-06c2911f2fda, 'name': SearchDatastore_Task, 'duration_secs': 0.027052} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.436730] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.437181] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 7188fa81-3a23-4f22-adb8-41a35b1af9a5/7188fa81-3a23-4f22-adb8-41a35b1af9a5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.437724] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8587a3a6-1ec9-4174-9737-6f9e689e3752 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.445225] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 944.445225] env[62558]: value = "task-1267088" [ 944.445225] env[62558]: _type = "Task" [ 944.445225] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.454264] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.483833] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updated VIF entry in instance network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.484234] env[62558]: DEBUG nova.network.neutron [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap37703c4e-05", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.636819] env[62558]: INFO nova.compute.manager [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Took 29.71 seconds to build instance. [ 944.694349] env[62558]: INFO nova.compute.manager [-] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Took 1.23 seconds to deallocate network for instance. [ 944.913124] env[62558]: DEBUG nova.compute.utils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.914038] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.914368] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 944.957558] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457035} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.957558] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 7188fa81-3a23-4f22-adb8-41a35b1af9a5/7188fa81-3a23-4f22-adb8-41a35b1af9a5.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.957558] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.957558] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10335167-a441-408c-beb8-49faf33c837e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.963087] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 944.963087] env[62558]: value = "task-1267089" [ 944.963087] env[62558]: _type = "Task" [ 944.963087] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.976167] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.987139] env[62558]: DEBUG oslo_concurrency.lockutils [req-6122485c-4299-41e4-8e9a-7d58b2b5da37 req-26a19430-9a39-4fca-85a4-be70eb76a88c service nova] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.995684] env[62558]: DEBUG nova.policy [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e6e2f1bae4d4bb68070770dbbbe5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78b69738b90142a4943069006e349445', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.117910] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.138517] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1da7f288-302a-442a-a5f4-9adda402d0af tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.228s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.200737] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.415958] env[62558]: DEBUG oslo_concurrency.lockutils [None req-eaf36a7a-c368-4ab7-a46c-76b9ba394ed7 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.417162] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.299s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.418761] env[62558]: INFO nova.compute.claims [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.421622] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.479000] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068063} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.479350] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.480423] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008f5e7c-949d-4cfd-83e3-b111251631f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.507226] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 7188fa81-3a23-4f22-adb8-41a35b1af9a5/7188fa81-3a23-4f22-adb8-41a35b1af9a5.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.507794] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0b8891d-0c64-4b60-b31b-6f7b7e1e9bf1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.531909] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 945.531909] env[62558]: value = "task-1267090" [ 945.531909] env[62558]: _type = "Task" [ 945.531909] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.544360] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267090, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.643830] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Successfully created port: 94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.042016] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267090, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.238202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.238202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.238202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.238202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.238202] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.240782] env[62558]: INFO nova.compute.manager [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Terminating instance [ 946.242882] env[62558]: DEBUG nova.compute.manager [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.243149] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.244465] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c968ff-36a7-4e6b-a5ec-825526ff0c25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.253307] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.253609] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e18853d-e918-4ff6-94d5-08ee5b142100 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.260924] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 946.260924] env[62558]: value = "task-1267091" [ 946.260924] env[62558]: _type = "Task" [ 946.260924] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.269375] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.435483] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.464452] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.464746] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.464932] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.465239] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.465474] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.465684] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.465987] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.466212] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.466451] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.466678] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.466916] env[62558]: DEBUG nova.virt.hardware [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.469767] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a5d9c3-eca7-4c37-beb9-e46cc2b3055d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.479547] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126a4e5e-cb6b-48f3-898c-4ebcbc681958 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.512965] env[62558]: DEBUG nova.compute.manager [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.513226] env[62558]: DEBUG nova.compute.manager [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing instance network info cache due to event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.513545] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.513697] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.513917] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.545658] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267090, 'name': ReconfigVM_Task, 'duration_secs': 0.534356} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.545943] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 7188fa81-3a23-4f22-adb8-41a35b1af9a5/7188fa81-3a23-4f22-adb8-41a35b1af9a5.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.547054] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-954e41a7-5e3e-434f-98fb-22faa7058ddc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.556088] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 946.556088] env[62558]: value = "task-1267092" [ 946.556088] env[62558]: _type = "Task" [ 946.556088] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.565120] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267092, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.725607] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ef0b3d16-704d-4435-9c23-a258d94a9983" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.725607] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.766156] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ae0c36-316b-4753-92ed-97d04510086e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.774422] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267091, 'name': PowerOffVM_Task, 'duration_secs': 0.251048} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.776362] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.776548] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.776837] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2df2101-3358-47a9-92d0-42a5a7edce76 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.778882] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33dc489-09a8-4b5d-b1ff-0bc57bffde14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.808437] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382bab74-6a75-4727-8a7b-f7c26364760f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.815776] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddb9f2b-4375-4c75-9b8e-809fd23dc3a1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.828662] env[62558]: DEBUG nova.compute.provider_tree [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.867960] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.868235] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.868395] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleting the datastore file [datastore1] a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.868738] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2402be55-7d68-4025-9ce8-46dc22264e35 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.875510] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 946.875510] env[62558]: value = "task-1267094" [ 946.875510] env[62558]: _type = "Task" [ 946.875510] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.884870] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.066147] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267092, 'name': Rename_Task, 'duration_secs': 0.17103} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.066440] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.066694] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-482b9ee7-2f02-4d1e-9494-92da4e2a71be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.072748] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 947.072748] env[62558]: value = "task-1267095" [ 947.072748] env[62558]: _type = "Task" [ 947.072748] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.080237] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.132614] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Successfully updated port: 94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.143177] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 947.265974] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updated VIF entry in instance network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.266365] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.332107] env[62558]: DEBUG nova.scheduler.client.report [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.385688] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267094, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.582607] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267095, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.635119] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.635249] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.635340] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.665889] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.769701] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.769835] env[62558]: DEBUG nova.compute.manager [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.769931] env[62558]: DEBUG nova.compute.manager [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing instance network info cache due to event network-changed-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.770134] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Acquiring lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.770283] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Acquired lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.770449] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Refreshing network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.839219] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.839219] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.840705] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.934s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.840977] env[62558]: DEBUG nova.objects.instance [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'resources' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.886145] env[62558]: DEBUG oslo_vmware.api [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267094, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.603144} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.886407] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.886621] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 947.886807] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.886981] env[62558]: INFO nova.compute.manager [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Took 1.64 seconds to destroy the instance on the hypervisor. [ 947.887248] env[62558]: DEBUG oslo.service.loopingcall [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.887440] env[62558]: DEBUG nova.compute.manager [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.887552] env[62558]: DEBUG nova.network.neutron [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 948.086710] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267095, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.166347] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.344356] env[62558]: DEBUG nova.compute.utils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.345779] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.346029] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.349393] env[62558]: DEBUG nova.objects.instance [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'numa_topology' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.352868] env[62558]: DEBUG nova.network.neutron [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating instance_info_cache with network_info: [{"id": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "address": "fa:16:3e:62:de:e5", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94d89474-fb", "ovs_interfaceid": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.396920] env[62558]: DEBUG nova.compute.manager [req-2e4ee57e-4b3e-4ab8-bd67-52810a28e850 req-d85ed56a-c482-48e4-9007-bcc738288cf2 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Received event network-vif-deleted-bdbc998f-cabe-438b-b3ce-6477b78a13e7 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.397149] env[62558]: INFO nova.compute.manager [req-2e4ee57e-4b3e-4ab8-bd67-52810a28e850 req-d85ed56a-c482-48e4-9007-bcc738288cf2 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Neutron deleted interface bdbc998f-cabe-438b-b3ce-6477b78a13e7; detaching it from the instance and deleting it from the info cache [ 948.397327] env[62558]: DEBUG nova.network.neutron [req-2e4ee57e-4b3e-4ab8-bd67-52810a28e850 req-d85ed56a-c482-48e4-9007-bcc738288cf2 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.401271] env[62558]: DEBUG nova.policy [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c5503a5bf734483af68683a8c853a71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9d4de9f22ec414d90eb8c2ed9c42d18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.515056] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updated VIF entry in instance network info cache for port 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.515448] env[62558]: DEBUG nova.network.neutron [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [{"id": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "address": "fa:16:3e:66:33:e8", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dba3d4c-c3", "ovs_interfaceid": "7dba3d4c-c3d7-4a6b-b03b-df7f321245c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.545497] env[62558]: DEBUG nova.compute.manager [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Received event network-vif-plugged-94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.546106] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.546331] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.546509] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.546810] env[62558]: DEBUG nova.compute.manager [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] No waiting events found dispatching network-vif-plugged-94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 948.546911] env[62558]: WARNING nova.compute.manager [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Received unexpected event network-vif-plugged-94d89474-fbcc-4208-aa16-91863dd3e2e5 for instance with vm_state building and task_state spawning. [ 948.547030] env[62558]: DEBUG nova.compute.manager [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Received event network-changed-94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.547218] env[62558]: DEBUG nova.compute.manager [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Refreshing instance network info cache due to event network-changed-94d89474-fbcc-4208-aa16-91863dd3e2e5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 948.547405] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Acquiring lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.585815] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267095, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.633931] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Successfully created port: 46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.852733] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.859901] env[62558]: DEBUG nova.objects.base [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Object Instance<49a58b46-207f-4515-b313-afcdb2d1ced3> lazy-loaded attributes: resources,numa_topology {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 948.863291] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.863599] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance network_info: |[{"id": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "address": "fa:16:3e:62:de:e5", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94d89474-fb", "ovs_interfaceid": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 948.863904] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Acquired lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.864089] env[62558]: DEBUG nova.network.neutron [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Refreshing network info cache for port 94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.865326] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:de:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94d89474-fbcc-4208-aa16-91863dd3e2e5', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.880390] env[62558]: DEBUG oslo.service.loopingcall [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.880778] env[62558]: DEBUG nova.network.neutron [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.881864] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.882503] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e81bbd7-c053-4db3-ac45-ed88b4003dfe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.910258] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a4732f9-769a-43b7-805f-563b85a47159 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.914985] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.914985] env[62558]: value = "task-1267096" [ 948.914985] env[62558]: _type = "Task" [ 948.914985] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.928593] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb20a291-9f79-438a-a031-22cc98801874 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.944905] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267096, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.973895] env[62558]: DEBUG nova.compute.manager [req-2e4ee57e-4b3e-4ab8-bd67-52810a28e850 req-d85ed56a-c482-48e4-9007-bcc738288cf2 service nova] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Detach interface failed, port_id=bdbc998f-cabe-438b-b3ce-6477b78a13e7, reason: Instance a1d242e6-1561-4bd4-8e39-281ab6346661 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.017771] env[62558]: DEBUG oslo_concurrency.lockutils [req-38ae1aa4-5328-4ff0-ac53-83b49e5763ae req-01eb3e61-b609-433b-9f54-1b379d80e8b8 service nova] Releasing lock "refresh_cache-4815ba3f-265f-466a-9850-4c325cdb88de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.085803] env[62558]: DEBUG oslo_vmware.api [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267095, 'name': PowerOnVM_Task, 'duration_secs': 1.896488} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.086099] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.086310] env[62558]: INFO nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Took 9.07 seconds to spawn the instance on the hypervisor. [ 949.086609] env[62558]: DEBUG nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.087405] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3759f90-0737-495f-9611-a8c3e702d840 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.260658] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b1acaa-e720-4ce3-a72e-1d0fa3101588 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.268396] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecac8e0-f354-492f-add9-170c75823664 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.298597] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf0dc52-fb27-401b-b3cd-5528c49e252e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.306181] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e8c614-2a2c-4c4f-a8c1-a34c433b0e41 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.319353] env[62558]: DEBUG nova.compute.provider_tree [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.383567] env[62558]: INFO nova.compute.manager [-] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Took 1.50 seconds to deallocate network for instance. [ 949.427849] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267096, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.609858] env[62558]: INFO nova.compute.manager [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Took 33.24 seconds to build instance. [ 949.611753] env[62558]: DEBUG nova.network.neutron [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updated VIF entry in instance network info cache for port 94d89474-fbcc-4208-aa16-91863dd3e2e5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.611753] env[62558]: DEBUG nova.network.neutron [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating instance_info_cache with network_info: [{"id": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "address": "fa:16:3e:62:de:e5", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94d89474-fb", "ovs_interfaceid": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.822417] env[62558]: DEBUG nova.scheduler.client.report [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.868609] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.890198] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.896644] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.896892] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.897062] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.897248] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.897397] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.898433] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.898941] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.898941] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.899229] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.899229] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.899574] env[62558]: DEBUG nova.virt.hardware [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.900783] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f24e3df-f90c-47e6-acbf-bda067652f73 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.909339] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc71bd6b-9380-443d-bf7a-96fe4072ceaa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.932450] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267096, 'name': CreateVM_Task, 'duration_secs': 0.566166} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.932649] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.933388] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.933699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.933997] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 949.934310] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e9f66f7-d375-48db-a48d-b09417c5e2ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.939486] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 949.939486] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52267240-f0ec-7aae-2c26-0f933d8328d7" [ 949.939486] env[62558]: _type = "Task" [ 949.939486] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.950660] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52267240-f0ec-7aae-2c26-0f933d8328d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.113381] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e079768-0c1e-49b8-bc78-263fcf3d7690 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.754s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.114021] env[62558]: DEBUG oslo_concurrency.lockutils [req-ecae6556-47d6-467b-8b87-b923ea062fe4 req-9c5372c5-400f-419a-a489-c514759e43ce service nova] Releasing lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.330994] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.490s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.336587] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.317s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.336587] env[62558]: DEBUG nova.objects.instance [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lazy-loading 'resources' on Instance uuid f8d6df86-52b8-4692-add4-1ffec82cc598 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.447806] env[62558]: DEBUG nova.compute.manager [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Received event network-vif-plugged-46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.448333] env[62558]: DEBUG oslo_concurrency.lockutils [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.448333] env[62558]: DEBUG oslo_concurrency.lockutils [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.448445] env[62558]: DEBUG oslo_concurrency.lockutils [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.448604] env[62558]: DEBUG nova.compute.manager [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] No waiting events found dispatching network-vif-plugged-46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.448771] env[62558]: WARNING nova.compute.manager [req-e220a352-f41a-4d88-85bf-5367c54a3e6d req-1701bb96-cc36-488f-acf0-900f1fedf99f service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Received unexpected event network-vif-plugged-46f92558-3c85-4803-aad8-7ba3157e3f67 for instance with vm_state building and task_state spawning. [ 950.453444] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52267240-f0ec-7aae-2c26-0f933d8328d7, 'name': SearchDatastore_Task, 'duration_secs': 0.011659} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.453647] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.453950] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.454182] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.454347] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.454573] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.454855] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23d5416f-de92-4929-be35-3cea00ba5e87 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.463821] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.464059] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.464799] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7491e664-abfc-4527-bbf7-2c83cf7eb19a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.470843] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 950.470843] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526edf94-eda8-8010-6eff-4fb281a2a699" [ 950.470843] env[62558]: _type = "Task" [ 950.470843] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.479328] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526edf94-eda8-8010-6eff-4fb281a2a699, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.532512] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Successfully updated port: 46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.843038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e76d237-cd3f-4a36-9902-16985dd4f11e tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 28.712s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.844052] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.726s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.844052] env[62558]: INFO nova.compute.manager [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Unshelving [ 950.983458] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526edf94-eda8-8010-6eff-4fb281a2a699, 'name': SearchDatastore_Task, 'duration_secs': 0.008686} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.984315] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e03a7d0-96d6-4731-a217-c129ebe3f815 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.990356] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 950.990356] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52102c07-87de-8707-c47e-a1bbe42d79c2" [ 950.990356] env[62558]: _type = "Task" [ 950.990356] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.002762] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52102c07-87de-8707-c47e-a1bbe42d79c2, 'name': SearchDatastore_Task, 'duration_secs': 0.008044} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.003037] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.003306] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.003594] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e78e2c0-73bb-4567-ac5a-5c1ce9dffc95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.010216] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 951.010216] env[62558]: value = "task-1267097" [ 951.010216] env[62558]: _type = "Task" [ 951.010216] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.019257] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267097, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.034847] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.034979] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.035140] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.119978] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7be47f-0224-45fc-a583-b7e8e941fa6f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.128160] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147cfb7f-d340-452b-a01d-d0f14060bc6c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.159782] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575f2226-1d9f-490e-a7db-4a7512450966 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.170523] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf4729e-89d3-48c0-855d-1d7af4de6c9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.188257] env[62558]: DEBUG nova.compute.provider_tree [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.200442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.200715] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.200939] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.201143] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.201603] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.206229] env[62558]: INFO nova.compute.manager [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Terminating instance [ 951.207615] env[62558]: DEBUG nova.compute.manager [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 951.207888] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.208710] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a05c93-347a-4c00-86e2-504d4aaeadaf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.219057] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.219328] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a68273a1-18ae-417c-a2c2-929a3c2ca05b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.227852] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 951.227852] env[62558]: value = "task-1267098" [ 951.227852] env[62558]: _type = "Task" [ 951.227852] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.238352] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.522503] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267097, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.569811] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.691798] env[62558]: DEBUG nova.scheduler.client.report [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.718326] env[62558]: DEBUG nova.network.neutron [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.741048] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267098, 'name': PowerOffVM_Task, 'duration_secs': 0.205019} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.741460] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.741749] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.742151] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21da0d3d-74be-4c2e-bbcf-fcd8d784b032 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.798696] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.798947] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.799152] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Deleting the datastore file [datastore2] 7188fa81-3a23-4f22-adb8-41a35b1af9a5 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.799435] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc994e04-6eca-4037-87ce-d0638e1c6a54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.805860] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for the task: (returnval){ [ 951.805860] env[62558]: value = "task-1267100" [ 951.805860] env[62558]: _type = "Task" [ 951.805860] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.814191] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.869590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.020658] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267097, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.197065] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.199629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.130s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.199893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.201988] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.885s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.203730] env[62558]: INFO nova.compute.claims [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.218813] env[62558]: INFO nova.scheduler.client.report [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleted allocations for instance f8d6df86-52b8-4692-add4-1ffec82cc598 [ 952.222971] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.222971] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Instance network_info: |[{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.224423] env[62558]: INFO nova.scheduler.client.report [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocations for instance 63050323-047e-4d73-91ae-859467b4b5a7 [ 952.226122] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:e8:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f256cfee-512d-4192-9aca-6750fdb1cd4c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46f92558-3c85-4803-aad8-7ba3157e3f67', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.233119] env[62558]: DEBUG oslo.service.loopingcall [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.235608] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.236173] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc242826-6cad-4dca-a8bc-9f5d18d12886 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.257026] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.257026] env[62558]: value = "task-1267101" [ 952.257026] env[62558]: _type = "Task" [ 952.257026] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.266103] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267101, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.317445] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.475262] env[62558]: DEBUG nova.compute.manager [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Received event network-changed-46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.475707] env[62558]: DEBUG nova.compute.manager [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Refreshing instance network info cache due to event network-changed-46f92558-3c85-4803-aad8-7ba3157e3f67. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.476087] env[62558]: DEBUG oslo_concurrency.lockutils [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] Acquiring lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.476351] env[62558]: DEBUG oslo_concurrency.lockutils [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] Acquired lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.476659] env[62558]: DEBUG nova.network.neutron [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Refreshing network info cache for port 46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.522858] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267097, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.482035} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.523273] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.523426] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.524028] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fcda609-5ade-48db-84af-22faca92d3e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.530316] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 952.530316] env[62558]: value = "task-1267102" [ 952.530316] env[62558]: _type = "Task" [ 952.530316] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.540265] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.727610] env[62558]: DEBUG oslo_concurrency.lockutils [None req-51aaf9fb-c6c5-4edb-8de4-a76de768cf08 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f8d6df86-52b8-4692-add4-1ffec82cc598" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.947s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.740049] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7821a542-b92f-40a2-8a60-bdfd6e0f0704 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "63050323-047e-4d73-91ae-859467b4b5a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.233s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.774546] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267101, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.818145] env[62558]: DEBUG oslo_vmware.api [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Task: {'id': task-1267100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.709451} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.818420] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.818612] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.818793] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.818972] env[62558]: INFO nova.compute.manager [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Took 1.61 seconds to destroy the instance on the hypervisor. [ 952.819242] env[62558]: DEBUG oslo.service.loopingcall [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.819444] env[62558]: DEBUG nova.compute.manager [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 952.819539] env[62558]: DEBUG nova.network.neutron [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.042641] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.315936} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.043023] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.044156] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddc968a-3038-4dc8-8b1a-7e14f0d48288 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.073047] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.076364] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a00b5ff7-0638-44bc-a4fa-97138d735af2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.103013] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 953.103013] env[62558]: value = "task-1267103" [ 953.103013] env[62558]: _type = "Task" [ 953.103013] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.112386] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267103, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.240820] env[62558]: DEBUG nova.network.neutron [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updated VIF entry in instance network info cache for port 46f92558-3c85-4803-aad8-7ba3157e3f67. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.241279] env[62558]: DEBUG nova.network.neutron [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.272282] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267101, 'name': CreateVM_Task, 'duration_secs': 0.77829} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.274884] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.275779] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.275968] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.276325] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.276588] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40bee297-577e-4823-93e6-b00053f9d554 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.281181] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 953.281181] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f2efe3-6025-72c2-0b85-dc96510e1934" [ 953.281181] env[62558]: _type = "Task" [ 953.281181] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.293080] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f2efe3-6025-72c2-0b85-dc96510e1934, 'name': SearchDatastore_Task, 'duration_secs': 0.008754} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.293429] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.293696] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.293978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.294252] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.294361] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.294692] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-543adb69-9624-420a-84ae-4939b135da69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.302533] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.302947] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.303734] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2de89ca0-4f90-430c-9ae2-885041896673 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.309862] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 953.309862] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d71baa-5e97-3cbb-7f97-bf99caa5db27" [ 953.309862] env[62558]: _type = "Task" [ 953.309862] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.325242] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52d71baa-5e97-3cbb-7f97-bf99caa5db27, 'name': SearchDatastore_Task, 'duration_secs': 0.008396} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.329008] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75ea44b6-63b1-4877-8304-e6bf509761aa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.334542] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 953.334542] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52316c81-12b4-8da0-b182-92bf059e246b" [ 953.334542] env[62558]: _type = "Task" [ 953.334542] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.351219] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52316c81-12b4-8da0-b182-92bf059e246b, 'name': SearchDatastore_Task, 'duration_secs': 0.012788} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.352426] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.352426] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.352720] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89edc1d6-c791-43d1-a3f0-27140f1b8eb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.360486] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 953.360486] env[62558]: value = "task-1267104" [ 953.360486] env[62558]: _type = "Task" [ 953.360486] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.370416] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267104, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.477131] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd233df5-e342-4920-a28a-b39bf87de7f2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.484940] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb3decb-5f7a-4d54-b901-d015c7cda47a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.515697] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9a5a60-7c6d-4322-b6f2-82c0b70a602f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.524335] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b55e4ec-a219-4640-a985-7767248e972f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.538342] env[62558]: DEBUG nova.compute.provider_tree [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.597224] env[62558]: DEBUG nova.network.neutron [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.616042] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267103, 'name': ReconfigVM_Task, 'duration_secs': 0.296827} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.616517] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.616961] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6cf2af9-fabf-41b2-a383-e0c0423ca9e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.625117] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 953.625117] env[62558]: value = "task-1267105" [ 953.625117] env[62558]: _type = "Task" [ 953.625117] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.641791] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267105, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.744144] env[62558]: DEBUG oslo_concurrency.lockutils [req-ed6ca285-e42f-4065-afb0-a548a734e125 req-4be1c3cb-b674-489e-a4e6-fae784dade86 service nova] Releasing lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.843796] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f642911c-bb9e-4187-8e01-65c45cb6e793" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.844144] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.844388] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "f642911c-bb9e-4187-8e01-65c45cb6e793-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.844643] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.845127] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.847148] env[62558]: INFO nova.compute.manager [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Terminating instance [ 953.848762] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "refresh_cache-f642911c-bb9e-4187-8e01-65c45cb6e793" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.848919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquired lock "refresh_cache-f642911c-bb9e-4187-8e01-65c45cb6e793" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.849095] env[62558]: DEBUG nova.network.neutron [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.869854] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267104, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411801} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.870105] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.870315] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.870530] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef0bebf3-bb84-48c3-b36f-864db296abae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.877605] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 953.877605] env[62558]: value = "task-1267106" [ 953.877605] env[62558]: _type = "Task" [ 953.877605] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.886798] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267106, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.041893] env[62558]: DEBUG nova.scheduler.client.report [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.100246] env[62558]: INFO nova.compute.manager [-] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Took 1.28 seconds to deallocate network for instance. [ 954.135448] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267105, 'name': Rename_Task, 'duration_secs': 0.206136} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.136313] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.136574] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51711ca7-651a-47e8-bb3d-18eba92e5786 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.142349] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 954.142349] env[62558]: value = "task-1267107" [ 954.142349] env[62558]: _type = "Task" [ 954.142349] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.149909] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.367687] env[62558]: DEBUG nova.network.neutron [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 954.390706] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267106, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067772} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.391021] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.391796] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d4070c-fc9d-4d4a-84ba-01d12d49e777 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.414434] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.414762] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f88ce0a0-2d12-47eb-9904-2fbf3eb61675 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.429413] env[62558]: DEBUG nova.network.neutron [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.436592] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 954.436592] env[62558]: value = "task-1267108" [ 954.436592] env[62558]: _type = "Task" [ 954.436592] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.446167] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267108, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.500390] env[62558]: DEBUG nova.compute.manager [req-c9fe3a28-3eb8-4341-ac5b-3c8af441b32e req-88c3bc3c-0362-4ff6-a93b-bc9d2e9c2446 service nova] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Received event network-vif-deleted-253ac594-cda6-41b3-a77e-c8ce05524fbd {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.547350] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.547961] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.550785] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.350s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.551058] env[62558]: DEBUG nova.objects.instance [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lazy-loading 'resources' on Instance uuid 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.606996] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.621016] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.621320] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.621501] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 954.652725] env[62558]: DEBUG oslo_vmware.api [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267107, 'name': PowerOnVM_Task, 'duration_secs': 0.432452} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.653019] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.653243] env[62558]: INFO nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Took 8.22 seconds to spawn the instance on the hypervisor. [ 954.653444] env[62558]: DEBUG nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.654222] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9d06ce-8be9-4b6d-b352-849bb402a75a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.932168] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Releasing lock "refresh_cache-f642911c-bb9e-4187-8e01-65c45cb6e793" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.932634] env[62558]: DEBUG nova.compute.manager [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.932834] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.933807] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef750a99-71a9-47d1-b854-1f28b6f50bfa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.949103] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.949409] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267108, 'name': ReconfigVM_Task, 'duration_secs': 0.295027} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.949631] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa6e0b66-5001-4b74-accc-bff93c20a4fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.951292] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfigured VM instance instance-0000005e to attach disk [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.951727] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2682cbb3-4e34-4107-91d3-5831a5217901 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.957940] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 954.957940] env[62558]: value = "task-1267110" [ 954.957940] env[62558]: _type = "Task" [ 954.957940] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.959147] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 954.959147] env[62558]: value = "task-1267109" [ 954.959147] env[62558]: _type = "Task" [ 954.959147] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.969359] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267110, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.972292] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267109, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.057223] env[62558]: DEBUG nova.compute.utils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.059026] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.059026] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.100635] env[62558]: DEBUG nova.policy [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.171075] env[62558]: INFO nova.compute.manager [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Took 19.86 seconds to build instance. [ 955.182838] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.183460] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.184233] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 955.326909] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0efc53-17c5-448f-9b35-c135bf1f3819 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.340280] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f233b3-4627-44cc-90ec-99ce2550b556 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.376419] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Successfully created port: 7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.379396] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49c22fe-4e34-421c-807f-10d50dccb038 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.387134] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd761876-b401-4d42-9bca-bd5adddc1fb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.400789] env[62558]: DEBUG nova.compute.provider_tree [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.473965] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267110, 'name': Rename_Task, 'duration_secs': 0.152609} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.477428] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.477723] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267109, 'name': PowerOffVM_Task, 'duration_secs': 0.183728} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.477929] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d88e9d9-16d1-4101-918f-d509e6d62d3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.479528] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.479707] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.479938] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f228374f-b767-47a7-bd08-00db08f8ba49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.487577] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 955.487577] env[62558]: value = "task-1267111" [ 955.487577] env[62558]: _type = "Task" [ 955.487577] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.494955] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267111, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.504063] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.504063] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.504250] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleting the datastore file [datastore2] f642911c-bb9e-4187-8e01-65c45cb6e793 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.506026] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef8e2d8a-82dc-4bce-9f90-87e170324b8a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.511225] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for the task: (returnval){ [ 955.511225] env[62558]: value = "task-1267113" [ 955.511225] env[62558]: _type = "Task" [ 955.511225] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.520651] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.561564] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.676580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f176bebc-8d51-447f-a6f2-dc7d6ff08948 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.379s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.905618] env[62558]: DEBUG nova.scheduler.client.report [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.000279] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267111, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.021288] env[62558]: DEBUG oslo_vmware.api [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Task: {'id': task-1267113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104867} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.021516] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.022033] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.022033] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.022122] env[62558]: INFO nova.compute.manager [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Took 1.09 seconds to destroy the instance on the hypervisor. [ 956.022321] env[62558]: DEBUG oslo.service.loopingcall [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.022517] env[62558]: DEBUG nova.compute.manager [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.022614] env[62558]: DEBUG nova.network.neutron [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.041513] env[62558]: DEBUG nova.network.neutron [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.411391] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.415242] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.749s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.416695] env[62558]: INFO nova.compute.claims [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.433308] env[62558]: INFO nova.scheduler.client.report [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted allocations for instance 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6 [ 956.451310] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [{"id": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "address": "fa:16:3e:3a:79:a0", "network": {"id": "dfbb553a-59d4-49c9-ac3d-0fd979b5df7e", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-682588595-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9d93976e6045c288ad6b152304ef72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9a83e75-1c", "ovs_interfaceid": "d9a83e75-1cf0-4747-90ee-2fec31fff5ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.498714] env[62558]: DEBUG oslo_vmware.api [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267111, 'name': PowerOnVM_Task, 'duration_secs': 0.518196} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.498992] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.499215] env[62558]: INFO nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Took 6.63 seconds to spawn the instance on the hypervisor. [ 956.499404] env[62558]: DEBUG nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.500184] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b2affd-9d6e-46c7-907d-5a432c8031c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.529582] env[62558]: DEBUG nova.compute.manager [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Received event network-changed-94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.529794] env[62558]: DEBUG nova.compute.manager [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Refreshing instance network info cache due to event network-changed-94d89474-fbcc-4208-aa16-91863dd3e2e5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.530053] env[62558]: DEBUG oslo_concurrency.lockutils [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] Acquiring lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.530182] env[62558]: DEBUG oslo_concurrency.lockutils [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] Acquired lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.530348] env[62558]: DEBUG nova.network.neutron [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Refreshing network info cache for port 94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 956.543787] env[62558]: DEBUG nova.network.neutron [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.574864] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.603674] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.603818] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.603867] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.604062] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.604219] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.604367] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.604582] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.604741] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.604906] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.605080] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.605256] env[62558]: DEBUG nova.virt.hardware [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.606335] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dedac3e-8e1b-4fb6-b49d-f2b0b341b0d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.613885] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b271f33f-257c-4d92-8921-c7f01a423484 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.808333] env[62558]: DEBUG nova.compute.manager [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Received event network-vif-plugged-7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.808491] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] Acquiring lock "591d326c-c20c-401b-a57b-895d59c3b418-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.808708] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] Lock "591d326c-c20c-401b-a57b-895d59c3b418-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.808878] env[62558]: DEBUG oslo_concurrency.lockutils [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] Lock "591d326c-c20c-401b-a57b-895d59c3b418-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.809329] env[62558]: DEBUG nova.compute.manager [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] No waiting events found dispatching network-vif-plugged-7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.809608] env[62558]: WARNING nova.compute.manager [req-cd4e110a-5cbf-4fae-a560-ca2b8dea5aa3 req-9c28740e-6cd7-4aef-b660-536bc9057c1e service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Received unexpected event network-vif-plugged-7e9fa3d9-6e23-4c06-b251-602a3c289637 for instance with vm_state building and task_state spawning. [ 956.943715] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b37159b1-21a8-41ff-b2cd-1a7156c3d241 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "42b2e70f-a46b-4984-b6d1-a2bfe348d1f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.672s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.946239] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Successfully updated port: 7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.954087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-2ac801d7-af70-46e0-88b3-02caee13497d" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.954087] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 956.954610] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.954836] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.955053] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.955282] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.955568] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.955642] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 956.955781] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 956.956186] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.015800] env[62558]: INFO nova.compute.manager [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Took 15.92 seconds to build instance. [ 957.046434] env[62558]: INFO nova.compute.manager [-] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Took 1.02 seconds to deallocate network for instance. [ 957.268111] env[62558]: DEBUG nova.network.neutron [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updated VIF entry in instance network info cache for port 94d89474-fbcc-4208-aa16-91863dd3e2e5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.268111] env[62558]: DEBUG nova.network.neutron [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating instance_info_cache with network_info: [{"id": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "address": "fa:16:3e:62:de:e5", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94d89474-fb", "ovs_interfaceid": "94d89474-fbcc-4208-aa16-91863dd3e2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.321275] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.321617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.321896] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.322177] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.322418] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.324916] env[62558]: INFO nova.compute.manager [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Terminating instance [ 957.327437] env[62558]: DEBUG nova.compute.manager [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.327704] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.328822] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42feaf9c-8760-4357-b595-8a7c016d7e54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.337031] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.337286] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb73c60f-4563-411f-8436-ff0232f57979 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.344143] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 957.344143] env[62558]: value = "task-1267114" [ 957.344143] env[62558]: _type = "Task" [ 957.344143] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.351843] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.451022] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.451197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.451351] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.460646] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.517406] env[62558]: DEBUG oslo_concurrency.lockutils [None req-222d06fd-1d7a-486b-9990-8e72f5db9d3c tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.434s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.552952] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.664777] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbd31c2-7038-4c8a-add1-837374029300 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.672843] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1253cdb4-deca-4199-8158-c77f38852cc7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.704838] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6f75d9-9944-419a-b39b-862e90bc8dec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.712534] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a80e044-1c49-4192-adcc-caf331573d22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.726158] env[62558]: DEBUG nova.compute.provider_tree [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.770387] env[62558]: DEBUG oslo_concurrency.lockutils [req-74dc8dac-affc-4bfc-87b1-2e86474a3bc4 req-7e47ed8e-e322-4a6a-8c74-9ee6b6a68d76 service nova] Releasing lock "refresh_cache-109f67db-21cf-40f9-966e-4730bff98bcf" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.855328] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267114, 'name': PowerOffVM_Task, 'duration_secs': 0.217655} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.855533] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.855743] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.856030] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09c6596a-2239-40f8-a124-a1406d365ca9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.915377] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.915492] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.915639] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleting the datastore file [datastore1] 0e5c1138-7f09-4b76-9c56-93c0ad947ad8 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.917996] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f2e9c9e-9c69-4aff-93a3-f52111f59271 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.921987] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 957.921987] env[62558]: value = "task-1267116" [ 957.921987] env[62558]: _type = "Task" [ 957.921987] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.929505] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267116, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.983119] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.983377] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.985940] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.140300] env[62558]: DEBUG nova.network.neutron [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Updating instance_info_cache with network_info: [{"id": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "address": "fa:16:3e:95:49:ed", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e9fa3d9-6e", "ovs_interfaceid": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.229186] env[62558]: DEBUG nova.scheduler.client.report [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.432127] env[62558]: DEBUG oslo_vmware.api [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267116, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41704} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.432236] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.432398] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.432580] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.432810] env[62558]: INFO nova.compute.manager [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 958.433057] env[62558]: DEBUG oslo.service.loopingcall [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.433266] env[62558]: DEBUG nova.compute.manager [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.433437] env[62558]: DEBUG nova.network.neutron [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.486044] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.643184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.643551] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Instance network_info: |[{"id": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "address": "fa:16:3e:95:49:ed", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e9fa3d9-6e", "ovs_interfaceid": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.644046] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:49:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f78b07ea-f425-4622-84f4-706a5d8820a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e9fa3d9-6e23-4c06-b251-602a3c289637', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.652077] env[62558]: DEBUG oslo.service.loopingcall [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.652336] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.652909] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4912e7a2-f696-489c-95e6-3f17ec427be9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.674454] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.674454] env[62558]: value = "task-1267117" [ 958.674454] env[62558]: _type = "Task" [ 958.674454] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.682358] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267117, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.734424] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.735019] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.738230] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.848s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.738452] env[62558]: DEBUG nova.objects.instance [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'resources' on Instance uuid a1d242e6-1561-4bd4-8e39-281ab6346661 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.771052] env[62558]: DEBUG nova.compute.manager [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 958.839291] env[62558]: DEBUG nova.compute.manager [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Received event network-changed-7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.839521] env[62558]: DEBUG nova.compute.manager [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Refreshing instance network info cache due to event network-changed-7e9fa3d9-6e23-4c06-b251-602a3c289637. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.839704] env[62558]: DEBUG oslo_concurrency.lockutils [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] Acquiring lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.839822] env[62558]: DEBUG oslo_concurrency.lockutils [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] Acquired lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.839980] env[62558]: DEBUG nova.network.neutron [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Refreshing network info cache for port 7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.011160] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.186098] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267117, 'name': CreateVM_Task, 'duration_secs': 0.502389} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.186322] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.187256] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.187256] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.187482] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.187961] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e69681c5-4e32-402d-b494-c6eed81df137 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.193073] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 959.193073] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bbc2ff-69c9-ebed-1314-e73f1cdbe8ca" [ 959.193073] env[62558]: _type = "Task" [ 959.193073] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.201265] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bbc2ff-69c9-ebed-1314-e73f1cdbe8ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.211731] env[62558]: DEBUG nova.network.neutron [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.241165] env[62558]: DEBUG nova.compute.utils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 959.245869] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 959.245869] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 959.285868] env[62558]: DEBUG nova.policy [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 959.292046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.561358] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f536d3f3-451f-4d90-90b0-9ac2454bc076 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.570898] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6454bf-7f43-4634-92c1-fa99b101f855 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.574886] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Successfully created port: 2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.579696] env[62558]: DEBUG nova.network.neutron [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Updated VIF entry in instance network info cache for port 7e9fa3d9-6e23-4c06-b251-602a3c289637. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.580066] env[62558]: DEBUG nova.network.neutron [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Updating instance_info_cache with network_info: [{"id": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "address": "fa:16:3e:95:49:ed", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e9fa3d9-6e", "ovs_interfaceid": "7e9fa3d9-6e23-4c06-b251-602a3c289637", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.606103] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb699a3-61cd-4572-a7c5-ff072ee6c010 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.615825] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206346f8-1ae3-4fef-872f-56511c640f22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.629009] env[62558]: DEBUG nova.compute.provider_tree [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.703548] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bbc2ff-69c9-ebed-1314-e73f1cdbe8ca, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.703935] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.704136] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.704380] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.704532] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.704772] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.704992] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd616d4e-0e0f-496c-99be-405de87492ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.713133] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.713320] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.714312] env[62558]: INFO nova.compute.manager [-] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Took 1.28 seconds to deallocate network for instance. [ 959.714543] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d862ad36-e278-41b2-b060-12652b228d31 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.722081] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 959.722081] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ddbd9-1e19-0389-8b9d-b2be627477b9" [ 959.722081] env[62558]: _type = "Task" [ 959.722081] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.729246] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ddbd9-1e19-0389-8b9d-b2be627477b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.749713] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.106293] env[62558]: DEBUG oslo_concurrency.lockutils [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] Releasing lock "refresh_cache-591d326c-c20c-401b-a57b-895d59c3b418" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.106645] env[62558]: DEBUG nova.compute.manager [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Received event network-vif-deleted-94ed9b4f-2b00-4433-8c7a-45cfc47135fc {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.106882] env[62558]: INFO nova.compute.manager [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Neutron deleted interface 94ed9b4f-2b00-4433-8c7a-45cfc47135fc; detaching it from the instance and deleting it from the info cache [ 960.107151] env[62558]: DEBUG nova.network.neutron [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.131554] env[62558]: DEBUG nova.scheduler.client.report [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.229238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.236025] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525ddbd9-1e19-0389-8b9d-b2be627477b9, 'name': SearchDatastore_Task, 'duration_secs': 0.025452} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.236025] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42a8bf20-73f0-4d14-be5f-edaf22bd19fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.241040] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 960.241040] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da945f-3fb9-889c-25e6-528126c8118a" [ 960.241040] env[62558]: _type = "Task" [ 960.241040] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.249077] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da945f-3fb9-889c-25e6-528126c8118a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.612984] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d1bb0ae-9e7b-48ff-b592-20f8e2442e3f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.620437] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e541b24-1729-46bc-abac-342ce1f072d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.635660] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.637877] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.768s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.638489] env[62558]: DEBUG nova.objects.instance [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'pci_requests' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.649891] env[62558]: DEBUG nova.compute.manager [req-72f67765-6d76-4682-8dc7-91bb3c4267ae req-3dce925d-6b9c-4d54-87c2-d0721570b449 service nova] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Detach interface failed, port_id=94ed9b4f-2b00-4433-8c7a-45cfc47135fc, reason: Instance 0e5c1138-7f09-4b76-9c56-93c0ad947ad8 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.661907] env[62558]: INFO nova.scheduler.client.report [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleted allocations for instance a1d242e6-1561-4bd4-8e39-281ab6346661 [ 960.751547] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da945f-3fb9-889c-25e6-528126c8118a, 'name': SearchDatastore_Task, 'duration_secs': 0.017449} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.751837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.752104] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 591d326c-c20c-401b-a57b-895d59c3b418/591d326c-c20c-401b-a57b-895d59c3b418.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.752396] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9395b49-a72f-475f-a98f-cfee973f7b78 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.759572] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.761488] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 960.761488] env[62558]: value = "task-1267118" [ 960.761488] env[62558]: _type = "Task" [ 960.761488] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.769806] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.785009] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.785281] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.785453] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.785642] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.785791] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.785959] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.786189] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.786352] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.786522] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.786690] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.786863] env[62558]: DEBUG nova.virt.hardware [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.787704] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e170d7-eb1e-4a47-8c41-9d24be08d501 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.795033] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1f591c-c3e9-45fe-ba64-1d34a1effccb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.967675] env[62558]: DEBUG nova.compute.manager [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Received event network-vif-plugged-2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.968696] env[62558]: DEBUG oslo_concurrency.lockutils [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] Acquiring lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.968696] env[62558]: DEBUG oslo_concurrency.lockutils [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.968696] env[62558]: DEBUG oslo_concurrency.lockutils [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.968696] env[62558]: DEBUG nova.compute.manager [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] No waiting events found dispatching network-vif-plugged-2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.968945] env[62558]: WARNING nova.compute.manager [req-566523e4-d498-4650-b739-3052d8b786ad req-545a96af-f237-4f52-9eb5-eb26450c4ce1 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Received unexpected event network-vif-plugged-2b9426f0-9075-4138-9397-479111c634a3 for instance with vm_state building and task_state spawning. [ 961.059148] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Successfully updated port: 2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.142439] env[62558]: DEBUG nova.objects.instance [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'numa_topology' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.168555] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9559a79e-3d62-4814-89d5-2659c20cc3d8 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "a1d242e6-1561-4bd4-8e39-281ab6346661" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.931s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.271655] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267118, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.562055] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.562055] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.562055] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.645179] env[62558]: INFO nova.compute.claims [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.771933] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.96181} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.772236] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 591d326c-c20c-401b-a57b-895d59c3b418/591d326c-c20c-401b-a57b-895d59c3b418.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.772451] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.772709] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d666d653-b643-499a-a7af-786a26b55e12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.779112] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 961.779112] env[62558]: value = "task-1267119" [ 961.779112] env[62558]: _type = "Task" [ 961.779112] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.787276] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.097438] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 962.229112] env[62558]: DEBUG nova.network.neutron [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Updating instance_info_cache with network_info: [{"id": "2b9426f0-9075-4138-9397-479111c634a3", "address": "fa:16:3e:52:31:9d", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9426f0-90", "ovs_interfaceid": "2b9426f0-9075-4138-9397-479111c634a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.289936] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117069} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.290248] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.291153] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7c5695-61c2-4837-8de1-e3e7ca3aea69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.316087] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 591d326c-c20c-401b-a57b-895d59c3b418/591d326c-c20c-401b-a57b-895d59c3b418.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.316758] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af157f84-fa74-41be-b456-233194ef0bf3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.337894] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 962.337894] env[62558]: value = "task-1267120" [ 962.337894] env[62558]: _type = "Task" [ 962.337894] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.346576] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.732915] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.733262] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Instance network_info: |[{"id": "2b9426f0-9075-4138-9397-479111c634a3", "address": "fa:16:3e:52:31:9d", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9426f0-90", "ovs_interfaceid": "2b9426f0-9075-4138-9397-479111c634a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.733764] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:31:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b9426f0-9075-4138-9397-479111c634a3', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.743388] env[62558]: DEBUG oslo.service.loopingcall [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.746320] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.746800] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-036696cc-f60a-4d3a-804b-79b79d2d2d3c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.771437] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.771437] env[62558]: value = "task-1267121" [ 962.771437] env[62558]: _type = "Task" [ 962.771437] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.782830] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267121, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.847503] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267120, 'name': ReconfigVM_Task, 'duration_secs': 0.312856} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.850191] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 591d326c-c20c-401b-a57b-895d59c3b418/591d326c-c20c-401b-a57b-895d59c3b418.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.851119] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc75b036-2572-49b5-b0fe-69b9b9de8bfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.859111] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 962.859111] env[62558]: value = "task-1267122" [ 962.859111] env[62558]: _type = "Task" [ 962.859111] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.874298] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267122, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.943981] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e45259f-75a2-41e1-b9aa-e953e7b11929 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.952914] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d735b8-9239-41d1-ae5a-05b2af5b6d86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.988925] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5341243-bc05-4514-954e-1bad5ceb4b1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.998637] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861bc6b0-3d78-4bfc-a1a5-9f98580cdd92 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.015472] env[62558]: DEBUG nova.compute.provider_tree [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.041290] env[62558]: DEBUG nova.compute.manager [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Received event network-changed-2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.041425] env[62558]: DEBUG nova.compute.manager [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Refreshing instance network info cache due to event network-changed-2b9426f0-9075-4138-9397-479111c634a3. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.042510] env[62558]: DEBUG oslo_concurrency.lockutils [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] Acquiring lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.042510] env[62558]: DEBUG oslo_concurrency.lockutils [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] Acquired lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.042510] env[62558]: DEBUG nova.network.neutron [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Refreshing network info cache for port 2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.294308] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267121, 'name': CreateVM_Task, 'duration_secs': 0.388033} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.294487] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 963.295218] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.295394] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.295870] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.295973] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bb81d85-b4aa-4767-92e7-2b3e23250668 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.302370] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 963.302370] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ccf45d-6a2e-c5e8-4b81-e359120857f2" [ 963.302370] env[62558]: _type = "Task" [ 963.302370] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.316281] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ccf45d-6a2e-c5e8-4b81-e359120857f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.369368] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267122, 'name': Rename_Task, 'duration_secs': 0.179888} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.369599] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.369865] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98d97f17-0683-4d8b-be52-575f8cb359ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.378531] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 963.378531] env[62558]: value = "task-1267123" [ 963.378531] env[62558]: _type = "Task" [ 963.378531] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.389427] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.519199] env[62558]: DEBUG nova.scheduler.client.report [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.743080] env[62558]: DEBUG nova.network.neutron [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Updated VIF entry in instance network info cache for port 2b9426f0-9075-4138-9397-479111c634a3. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.743793] env[62558]: DEBUG nova.network.neutron [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Updating instance_info_cache with network_info: [{"id": "2b9426f0-9075-4138-9397-479111c634a3", "address": "fa:16:3e:52:31:9d", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9426f0-90", "ovs_interfaceid": "2b9426f0-9075-4138-9397-479111c634a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.814610] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ccf45d-6a2e-c5e8-4b81-e359120857f2, 'name': SearchDatastore_Task, 'duration_secs': 0.011608} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.814962] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.815110] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.815272] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.815464] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.815608] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.815922] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03ad0bde-25f8-4b29-aeb5-47518de535fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.826530] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.827370] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.827764] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-155d10d3-8d08-486b-82df-35d7ec2666ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.834756] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 963.834756] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb179f-efbd-45ea-2792-1d6a29d25237" [ 963.834756] env[62558]: _type = "Task" [ 963.834756] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.843761] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb179f-efbd-45ea-2792-1d6a29d25237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.889763] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267123, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.901745] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.902121] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.024062] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.386s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.026909] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.420s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.027034] env[62558]: DEBUG nova.objects.instance [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lazy-loading 'resources' on Instance uuid 7188fa81-3a23-4f22-adb8-41a35b1af9a5 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.065354] env[62558]: INFO nova.network.neutron [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 964.248111] env[62558]: DEBUG oslo_concurrency.lockutils [req-6455a0d1-5e66-4964-be3d-947a04e09610 req-d9e10db7-acd8-4390-8ad6-9ff58f79eaf7 service nova] Releasing lock "refresh_cache-ef0b3d16-704d-4435-9c23-a258d94a9983" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.344862] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb179f-efbd-45ea-2792-1d6a29d25237, 'name': SearchDatastore_Task, 'duration_secs': 0.009235} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.345670] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f2aea40-061d-462e-b0d3-b3d111d1b1d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.350955] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 964.350955] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ea9453-2b65-da3f-2f3b-e7a42c854957" [ 964.350955] env[62558]: _type = "Task" [ 964.350955] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.358781] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ea9453-2b65-da3f-2f3b-e7a42c854957, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.389222] env[62558]: DEBUG oslo_vmware.api [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267123, 'name': PowerOnVM_Task, 'duration_secs': 0.635008} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.389510] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.389720] env[62558]: INFO nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Took 7.81 seconds to spawn the instance on the hypervisor. [ 964.389937] env[62558]: DEBUG nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.390753] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26b168f-91f9-45f8-8ddd-3c6632a75bd6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.408116] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 964.742875] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148fd45e-8434-4ebe-ba36-a4642c202347 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.751142] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53fa9c5-18f0-4004-80bf-dc073aef4cfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.780665] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17429d9f-1872-4626-829e-3c170fd1e750 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.787722] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efedb6f9-9015-41f7-970c-497763175aab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.800816] env[62558]: DEBUG nova.compute.provider_tree [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.862203] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ea9453-2b65-da3f-2f3b-e7a42c854957, 'name': SearchDatastore_Task, 'duration_secs': 0.009345} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.862493] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.862761] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ef0b3d16-704d-4435-9c23-a258d94a9983/ef0b3d16-704d-4435-9c23-a258d94a9983.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.863032] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c23dc8d-042e-4a80-8723-e235ff505e1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.870170] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 964.870170] env[62558]: value = "task-1267124" [ 964.870170] env[62558]: _type = "Task" [ 964.870170] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.879204] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.910372] env[62558]: INFO nova.compute.manager [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Took 21.61 seconds to build instance. [ 964.927678] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.304725] env[62558]: DEBUG nova.scheduler.client.report [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.383575] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267124, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.412191] env[62558]: DEBUG oslo_concurrency.lockutils [None req-900a3c81-a850-48df-98e8-100c28c942b1 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.132s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.489488] env[62558]: DEBUG nova.compute.manager [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.489719] env[62558]: DEBUG oslo_concurrency.lockutils [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.489928] env[62558]: DEBUG oslo_concurrency.lockutils [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.490494] env[62558]: DEBUG oslo_concurrency.lockutils [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.490709] env[62558]: DEBUG nova.compute.manager [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] No waiting events found dispatching network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.490882] env[62558]: WARNING nova.compute.manager [req-2a44a0db-1a87-409c-9e2e-7eb24ab52b93 req-7be6bddf-5f80-4f72-b136-4334595a4bcf service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received unexpected event network-vif-plugged-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a for instance with vm_state shelved_offloaded and task_state spawning. [ 965.621755] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.621948] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.622149] env[62558]: DEBUG nova.network.neutron [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.760729] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ec7513-37b6-48e8-a681-3e0cbe157cca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.768546] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Suspending the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 965.768792] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-76bf2049-05ad-45ad-bf19-2d6bb9863457 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.778086] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 965.778086] env[62558]: value = "task-1267125" [ 965.778086] env[62558]: _type = "Task" [ 965.778086] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.785148] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267125, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.808625] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.811287] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.350s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.811287] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.811482] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 965.811675] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.259s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.811928] env[62558]: DEBUG nova.objects.instance [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lazy-loading 'resources' on Instance uuid f642911c-bb9e-4187-8e01-65c45cb6e793 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.813772] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3617552b-f3eb-426c-957b-e1112019c8ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.823404] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee28ef3d-7bfe-4014-b427-042498d2baaa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.838882] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba4b7ee-3527-4382-a346-a3d33861abc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.847429] env[62558]: INFO nova.scheduler.client.report [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Deleted allocations for instance 7188fa81-3a23-4f22-adb8-41a35b1af9a5 [ 965.849403] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3ef3d5-3659-4f1e-beff-0b7a0f9360f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.887915] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179455MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 965.888508] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.899847] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267124, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.285889] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267125, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.362638] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5c769ced-e999-4eee-904f-69287b7914e7 tempest-ServerMetadataNegativeTestJSON-1372520468 tempest-ServerMetadataNegativeTestJSON-1372520468-project-member] Lock "7188fa81-3a23-4f22-adb8-41a35b1af9a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.161s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.387474] env[62558]: DEBUG nova.network.neutron [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.400931] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267124, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.600688] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8e7a03-e7cb-49d6-b1e0-a27bafe5f1c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.608471] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db65da4-c174-4403-99f7-afa0c2ce6e14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.639777] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ae418f-6aa8-4789-bb77-be75ee1bb99e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.648765] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f4528a-13e0-40c4-bd87-e615b75bee7d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.667028] env[62558]: DEBUG nova.compute.provider_tree [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.788545] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267125, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.889827] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.901845] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267124, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.909159} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.902283] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ef0b3d16-704d-4435-9c23-a258d94a9983/ef0b3d16-704d-4435-9c23-a258d94a9983.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 966.902567] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.902872] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea2b7443-b1d3-4c19-8b8b-6fce902a707c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.909955] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 966.909955] env[62558]: value = "task-1267126" [ 966.909955] env[62558]: _type = "Task" [ 966.909955] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.920145] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.922583] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a9dc9a8da60cb069f01b48e288bd053e',container_format='bare',created_at=2024-10-11T05:06:52Z,direct_url=,disk_format='vmdk',id=bf57715a-d9cb-4e6c-885f-8cbb16b14efd,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-446492939-shelved',owner='7663e884862d43b49eabe6396d8cd107',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-11T05:07:09Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.922852] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.923064] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.923260] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.923445] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.923699] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.923926] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.924110] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.924287] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.924453] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.924632] env[62558]: DEBUG nova.virt.hardware [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.925537] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0255980-03c8-49ed-9869-d29d0f3c725c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.933505] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d0d4a3-4035-4612-807d-34006bd66734 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.948644] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:85:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53ebf5df-5ecb-4a0c-a163-d88165639de0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.957076] env[62558]: DEBUG oslo.service.loopingcall [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.957476] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 966.957738] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-385d0881-b1ce-4ea7-8ffa-cac09db7b4be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.981071] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.981071] env[62558]: value = "task-1267127" [ 966.981071] env[62558]: _type = "Task" [ 966.981071] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.990468] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267127, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.174367] env[62558]: DEBUG nova.scheduler.client.report [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.289705] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267125, 'name': SuspendVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.357029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.357029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.421061] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158305} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.421061] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.421061] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0c5bda-8115-453d-9ea3-b52cb43d3761 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.447932] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] ef0b3d16-704d-4435-9c23-a258d94a9983/ef0b3d16-704d-4435-9c23-a258d94a9983.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.449721] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95292d2c-f004-4acc-89fa-b8920b8c530c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.473119] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 967.473119] env[62558]: value = "task-1267128" [ 967.473119] env[62558]: _type = "Task" [ 967.473119] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.485617] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267128, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.497187] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267127, 'name': CreateVM_Task, 'duration_secs': 0.391915} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.497290] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 967.497916] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.498117] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.498557] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.498838] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ed34a7e-2318-46f4-94c7-3e24c3aaa8fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.505127] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 967.505127] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524245c3-4f1b-d4b8-d847-e9fdf43c394b" [ 967.505127] env[62558]: _type = "Task" [ 967.505127] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.514021] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524245c3-4f1b-d4b8-d847-e9fdf43c394b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.524249] env[62558]: DEBUG nova.compute.manager [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.524503] env[62558]: DEBUG nova.compute.manager [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing instance network info cache due to event network-changed-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.524915] env[62558]: DEBUG oslo_concurrency.lockutils [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.524967] env[62558]: DEBUG oslo_concurrency.lockutils [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.525211] env[62558]: DEBUG nova.network.neutron [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Refreshing network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.680676] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.683117] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.672s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.685214] env[62558]: INFO nova.compute.claims [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.714017] env[62558]: INFO nova.scheduler.client.report [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Deleted allocations for instance f642911c-bb9e-4187-8e01-65c45cb6e793 [ 967.786522] env[62558]: DEBUG oslo_vmware.api [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267125, 'name': SuspendVM_Task, 'duration_secs': 1.686369} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.786744] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Suspended the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 967.786929] env[62558]: DEBUG nova.compute.manager [None req-794f3564-24a2-416d-acd0-94c5cd453521 tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.788256] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcff3722-e9ee-4243-8e5c-dec62a2f1ffc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.884592] env[62558]: DEBUG nova.compute.utils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.986062] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267128, 'name': ReconfigVM_Task, 'duration_secs': 0.318722} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.986391] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Reconfigured VM instance instance-00000060 to attach disk [datastore2] ef0b3d16-704d-4435-9c23-a258d94a9983/ef0b3d16-704d-4435-9c23-a258d94a9983.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.987064] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbc9892d-23cd-4ed2-92b3-d03d20390698 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.994784] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 967.994784] env[62558]: value = "task-1267129" [ 967.994784] env[62558]: _type = "Task" [ 967.994784] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.003735] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267129, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.016272] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.016543] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Processing image bf57715a-d9cb-4e6c-885f-8cbb16b14efd {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.016987] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.017137] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.017377] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.017738] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7eeb2194-23a9-4b8c-92b4-2e9111646e9b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.026327] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.026548] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.027318] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c1af981-415c-40fb-abb6-c3b9c412e1e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.033898] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 968.033898] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52107f92-4325-16fd-1876-f557500b6d24" [ 968.033898] env[62558]: _type = "Task" [ 968.033898] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.041263] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52107f92-4325-16fd-1876-f557500b6d24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.224114] env[62558]: DEBUG oslo_concurrency.lockutils [None req-53833c1b-dd76-413c-b2bb-757ed15676b0 tempest-ServerShowV247Test-266680753 tempest-ServerShowV247Test-266680753-project-member] Lock "f642911c-bb9e-4187-8e01-65c45cb6e793" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.380s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.267749] env[62558]: DEBUG nova.network.neutron [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updated VIF entry in instance network info cache for port 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.268135] env[62558]: DEBUG nova.network.neutron [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.364620] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.505229] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267129, 'name': Rename_Task, 'duration_secs': 0.141647} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.505527] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 968.505787] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7239eb2d-7b4d-4b91-85d9-20dfdbe2bee0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.512376] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 968.512376] env[62558]: value = "task-1267130" [ 968.512376] env[62558]: _type = "Task" [ 968.512376] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.520692] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.543857] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Preparing fetch location {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 968.545059] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Fetch image to [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c/OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c.vmdk {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 968.545059] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Downloading stream optimized image bf57715a-d9cb-4e6c-885f-8cbb16b14efd to [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c/OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c.vmdk on the data store datastore1 as vApp {{(pid=62558) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 968.545059] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Downloading image file data bf57715a-d9cb-4e6c-885f-8cbb16b14efd to the ESX as VM named 'OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c' {{(pid=62558) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 968.622033] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 968.622033] env[62558]: value = "resgroup-9" [ 968.622033] env[62558]: _type = "ResourcePool" [ 968.622033] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 968.622349] env[62558]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0dad11d5-3e5a-4df0-a053-47657df6c248 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.642834] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lease: (returnval){ [ 968.642834] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 968.642834] env[62558]: _type = "HttpNfcLease" [ 968.642834] env[62558]: } obtained for vApp import into resource pool (val){ [ 968.642834] env[62558]: value = "resgroup-9" [ 968.642834] env[62558]: _type = "ResourcePool" [ 968.642834] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 968.643094] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the lease: (returnval){ [ 968.643094] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 968.643094] env[62558]: _type = "HttpNfcLease" [ 968.643094] env[62558]: } to be ready. {{(pid=62558) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 968.649719] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 968.649719] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 968.649719] env[62558]: _type = "HttpNfcLease" [ 968.649719] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 968.770732] env[62558]: DEBUG oslo_concurrency.lockutils [req-b732ae49-f98b-4f46-b3bb-f2a830fdeac2 req-baf07099-123e-4846-ba0a-60627e516a2a service nova] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.932985] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c8ca64-85f3-4586-bc94-d2766b706c04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.944724] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80da8346-1d6f-4d80-b987-d68df051649e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.981176] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6879683d-bda0-4771-8ffb-e7542e5fe667 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.991299] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8000702b-38d2-477d-9f9e-26b3b3abb008 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.009570] env[62558]: DEBUG nova.compute.provider_tree [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.021517] env[62558]: DEBUG oslo_vmware.api [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267130, 'name': PowerOnVM_Task, 'duration_secs': 0.465851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.022032] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 969.022402] env[62558]: INFO nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Took 8.26 seconds to spawn the instance on the hypervisor. [ 969.022526] env[62558]: DEBUG nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.023371] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40f2990-59fd-477b-af97-c761d3937ab3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.152695] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 969.152695] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 969.152695] env[62558]: _type = "HttpNfcLease" [ 969.152695] env[62558]: } is initializing. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 969.443724] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.443724] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.443724] env[62558]: INFO nova.compute.manager [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Attaching volume 4d0add79-8ef3-4cf7-be38-7fd542cda129 to /dev/sdb [ 969.478219] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3a1c31-ea28-4a85-be98-71bfc3da5792 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.485332] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "591d326c-c20c-401b-a57b-895d59c3b418" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.485565] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.485763] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "591d326c-c20c-401b-a57b-895d59c3b418-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.485949] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.486136] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.488207] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b09f9f-154c-4bb1-b103-5e49ccb86211 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.490801] env[62558]: INFO nova.compute.manager [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Terminating instance [ 969.492695] env[62558]: DEBUG nova.compute.manager [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.492925] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.493628] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f276d9ee-372a-4afa-bc4a-2b1154102b1b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.502385] env[62558]: DEBUG nova.virt.block_device [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating existing volume attachment record: 4f2bcbf0-c2b8-4472-968f-b74342b5bc15 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 969.506092] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.506314] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb97c74a-15a6-44ed-99c9-89c127fa56db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.516905] env[62558]: DEBUG nova.scheduler.client.report [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.540951] env[62558]: INFO nova.compute.manager [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Took 21.89 seconds to build instance. [ 969.567260] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.567493] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.567684] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleting the datastore file [datastore2] 591d326c-c20c-401b-a57b-895d59c3b418 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.568173] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02fccd81-162f-4426-85fc-71f36eb92f26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.575622] env[62558]: DEBUG oslo_vmware.api [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 969.575622] env[62558]: value = "task-1267133" [ 969.575622] env[62558]: _type = "Task" [ 969.575622] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.582356] env[62558]: DEBUG oslo_vmware.api [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.652447] env[62558]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 969.652447] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 969.652447] env[62558]: _type = "HttpNfcLease" [ 969.652447] env[62558]: } is ready. {{(pid=62558) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 969.652755] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 969.652755] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520f9855-c29b-d76a-93a0-0d8098e7add9" [ 969.652755] env[62558]: _type = "HttpNfcLease" [ 969.652755] env[62558]: }. {{(pid=62558) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 969.653486] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1e2981-ac13-4875-b134-e80a1d908aa0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.660562] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk from lease info. {{(pid=62558) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 969.660823] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk. {{(pid=62558) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 969.728095] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c36401a4-bcda-4e35-8932-e2cd183119e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.021509] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.022231] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 970.027464] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.736s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.042199] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7be19765-aa56-42e7-ae70-ea27ee78260d tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.401s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.086449] env[62558]: DEBUG oslo_vmware.api [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275468} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.088215] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.089106] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.089106] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.089106] env[62558]: INFO nova.compute.manager [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Took 0.60 seconds to destroy the instance on the hypervisor. [ 970.089106] env[62558]: DEBUG oslo.service.loopingcall [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.089258] env[62558]: DEBUG nova.compute.manager [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.089285] env[62558]: DEBUG nova.network.neutron [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.382951] env[62558]: DEBUG nova.compute.manager [req-578d3f06-d3ca-4ffe-a6c1-a764faf5f878 req-c4d9271b-04df-4d9a-a767-338e9bb6ecf0 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Received event network-vif-deleted-7e9fa3d9-6e23-4c06-b251-602a3c289637 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.383104] env[62558]: INFO nova.compute.manager [req-578d3f06-d3ca-4ffe-a6c1-a764faf5f878 req-c4d9271b-04df-4d9a-a767-338e9bb6ecf0 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Neutron deleted interface 7e9fa3d9-6e23-4c06-b251-602a3c289637; detaching it from the instance and deleting it from the info cache [ 970.383312] env[62558]: DEBUG nova.network.neutron [req-578d3f06-d3ca-4ffe-a6c1-a764faf5f878 req-c4d9271b-04df-4d9a-a767-338e9bb6ecf0 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.531053] env[62558]: DEBUG nova.compute.utils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.535095] env[62558]: INFO nova.compute.claims [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.539100] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.539282] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.578472] env[62558]: DEBUG nova.policy [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61cb5cd30d434f1eb594dd4521dab2d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e1d3e1dcd64e918c1e436e691e290b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.852732] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Successfully created port: a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.857560] env[62558]: DEBUG nova.network.neutron [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.887374] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-006debc3-b552-4ab2-a538-dc47a0e46d98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.899392] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2755d9-4bbe-4e78-be3b-0e264c2edcb5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.935443] env[62558]: DEBUG nova.compute.manager [req-578d3f06-d3ca-4ffe-a6c1-a764faf5f878 req-c4d9271b-04df-4d9a-a767-338e9bb6ecf0 service nova] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Detach interface failed, port_id=7e9fa3d9-6e23-4c06-b251-602a3c289637, reason: Instance 591d326c-c20c-401b-a57b-895d59c3b418 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.943822] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Completed reading data from the image iterator. {{(pid=62558) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 970.944135] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 970.946931] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ad4731-03f1-45fd-8e77-df970c1e2d2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.951962] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk is in state: ready. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 970.952118] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk. {{(pid=62558) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 970.952293] env[62558]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-958ed793-d78f-46e9-8959-5d75a20fc576 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.992731] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ef0b3d16-704d-4435-9c23-a258d94a9983" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.993093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.993348] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.993588] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.993853] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.996212] env[62558]: INFO nova.compute.manager [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Terminating instance [ 970.998184] env[62558]: DEBUG nova.compute.manager [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.998435] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.999352] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de1c566-e0fc-4154-a5e0-bd74755a9539 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.007975] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.008314] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-070d95d2-6bd3-47dc-9103-ba78d2838140 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.014860] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 971.014860] env[62558]: value = "task-1267137" [ 971.014860] env[62558]: _type = "Task" [ 971.014860] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.026741] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.039779] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 971.045400] env[62558]: INFO nova.compute.resource_tracker [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating resource usage from migration 9b9fa3c1-4591-4035-8fc3-73041d59ebdd [ 971.226345] env[62558]: DEBUG oslo_vmware.rw_handles [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c3160b-a488-50a6-791e-b2de261db9e1/disk-0.vmdk. {{(pid=62558) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 971.226656] env[62558]: INFO nova.virt.vmwareapi.images [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Downloaded image file data bf57715a-d9cb-4e6c-885f-8cbb16b14efd [ 971.227571] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fa5415-a40a-4ba1-9e05-4fd3589c9e1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.246309] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d09d0871-70b0-45f9-a327-b57e683f2a41 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.276963] env[62558]: INFO nova.virt.vmwareapi.images [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] The imported VM was unregistered [ 971.279360] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Caching image {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 971.279619] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Creating directory with path [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.279901] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71184fdf-9120-4d23-9d29-c4ac82ec6400 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.292259] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Created directory with path [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.292465] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c/OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c.vmdk to [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk. {{(pid=62558) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 971.292739] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-82c5e448-b80b-466f-8afd-1f742439cfe0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.301507] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 971.301507] env[62558]: value = "task-1267139" [ 971.301507] env[62558]: _type = "Task" [ 971.301507] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.308767] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.335156] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd615828-5eee-493e-8b8e-8aafe2693a5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.344904] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4bdb97c-2d54-4231-a6bb-1dafdf6606f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.374872] env[62558]: INFO nova.compute.manager [-] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Took 1.29 seconds to deallocate network for instance. [ 971.377180] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4dbe79-95d3-41ea-985e-11e552f3ccbe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.387758] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9728eb6-e31b-4cb4-b533-6383c36f8711 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.404435] env[62558]: DEBUG nova.compute.provider_tree [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.525429] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267137, 'name': PowerOffVM_Task, 'duration_secs': 0.207819} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.525720] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.525890] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.526191] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c8d7dd9-70aa-4081-af4c-ce8103d6419a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.593692] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.593901] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.594071] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore2] ef0b3d16-704d-4435-9c23-a258d94a9983 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.595074] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0064f1d-4b1e-45ed-aa9f-8e1e5a191a81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.600944] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 971.600944] env[62558]: value = "task-1267141" [ 971.600944] env[62558]: _type = "Task" [ 971.600944] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.611741] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.814535] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.885092] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.907619] env[62558]: DEBUG nova.scheduler.client.report [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.055714] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 972.083710] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.084017] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.084183] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.084367] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.084515] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.084662] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.084881] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.085053] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.085232] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.085397] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.085571] env[62558]: DEBUG nova.virt.hardware [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.086517] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f706019f-0565-4fe9-8ec9-ff5a5591a64b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.097515] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00dbec58-7106-4ff5-8213-974f6589f882 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.119424] env[62558]: DEBUG oslo_vmware.api [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204116} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.119700] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.119889] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.120093] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.120282] env[62558]: INFO nova.compute.manager [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Took 1.12 seconds to destroy the instance on the hypervisor. [ 972.120534] env[62558]: DEBUG oslo.service.loopingcall [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.120742] env[62558]: DEBUG nova.compute.manager [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.120862] env[62558]: DEBUG nova.network.neutron [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.316518] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.320654] env[62558]: DEBUG nova.compute.manager [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Received event network-vif-plugged-a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.320885] env[62558]: DEBUG oslo_concurrency.lockutils [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] Acquiring lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.321165] env[62558]: DEBUG oslo_concurrency.lockutils [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.321286] env[62558]: DEBUG oslo_concurrency.lockutils [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.321457] env[62558]: DEBUG nova.compute.manager [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] No waiting events found dispatching network-vif-plugged-a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.321625] env[62558]: WARNING nova.compute.manager [req-747afc1d-3fea-4450-9f96-56b7cdfbf6fc req-36ab3b37-8125-48fa-98cb-fb4f0741b6d5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Received unexpected event network-vif-plugged-a703fd04-54f4-4a70-a8d2-8b588f0b5658 for instance with vm_state building and task_state spawning. [ 972.410899] env[62558]: DEBUG nova.compute.manager [req-3bdda34a-5e22-4911-8dda-41f879eaf270 req-76b659b6-38fa-42cd-ba61-ab41e5e4fb05 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Received event network-vif-deleted-2b9426f0-9075-4138-9397-479111c634a3 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.411155] env[62558]: INFO nova.compute.manager [req-3bdda34a-5e22-4911-8dda-41f879eaf270 req-76b659b6-38fa-42cd-ba61-ab41e5e4fb05 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Neutron deleted interface 2b9426f0-9075-4138-9397-479111c634a3; detaching it from the instance and deleting it from the info cache [ 972.411338] env[62558]: DEBUG nova.network.neutron [req-3bdda34a-5e22-4911-8dda-41f879eaf270 req-76b659b6-38fa-42cd-ba61-ab41e5e4fb05 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.413498] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.386s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.413699] env[62558]: INFO nova.compute.manager [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Migrating [ 972.420324] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.191s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.420579] env[62558]: DEBUG nova.objects.instance [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lazy-loading 'resources' on Instance uuid 0e5c1138-7f09-4b76-9c56-93c0ad947ad8 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.816820] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.884523] env[62558]: DEBUG nova.network.neutron [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.898382] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Successfully updated port: a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.922147] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ea786d0-0aa1-49ee-8e44-f962c5e39e3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.931318] env[62558]: DEBUG nova.compute.manager [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Received event network-changed-a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.931686] env[62558]: DEBUG nova.compute.manager [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Refreshing instance network info cache due to event network-changed-a703fd04-54f4-4a70-a8d2-8b588f0b5658. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 972.932076] env[62558]: DEBUG oslo_concurrency.lockutils [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] Acquiring lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.932362] env[62558]: DEBUG oslo_concurrency.lockutils [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] Acquired lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.932627] env[62558]: DEBUG nova.network.neutron [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Refreshing network info cache for port a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.935934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.935934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.936025] env[62558]: DEBUG nova.network.neutron [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.947461] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e75871-4fb9-442c-bfaa-e4bdc9478909 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.997631] env[62558]: DEBUG nova.compute.manager [req-3bdda34a-5e22-4911-8dda-41f879eaf270 req-76b659b6-38fa-42cd-ba61-ab41e5e4fb05 service nova] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Detach interface failed, port_id=2b9426f0-9075-4138-9397-479111c634a3, reason: Instance ef0b3d16-704d-4435-9c23-a258d94a9983 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 973.229884] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7cdce6-e345-4aa8-808b-a34f1c644adf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.240170] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b737c7ed-0ff0-4323-875f-c53635585d93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.276088] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d4d7a9-0b07-4a4e-bcdf-1a6a64ad078c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.286139] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e169b68-c76e-4351-89be-843fcf072a60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.302640] env[62558]: DEBUG nova.compute.provider_tree [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.316644] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.387467] env[62558]: INFO nova.compute.manager [-] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Took 1.27 seconds to deallocate network for instance. [ 973.401916] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.471496] env[62558]: DEBUG nova.network.neutron [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.558270] env[62558]: DEBUG nova.network.neutron [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.748561] env[62558]: DEBUG nova.network.neutron [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.806448] env[62558]: DEBUG nova.scheduler.client.report [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.820062] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267139, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.330165} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.820352] env[62558]: INFO nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c/OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c.vmdk to [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk. [ 973.820540] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Cleaning up location [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 973.820705] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_2e2274d2-afe4-4650-9ed3-5613bfc9328c {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.821548] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1b79a2ad-767c-42d3-a994-6ab921f1919d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.828042] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 973.828042] env[62558]: value = "task-1267143" [ 973.828042] env[62558]: _type = "Task" [ 973.828042] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.836129] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.894598] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.991025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.991465] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.994100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.994100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.994100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.995335] env[62558]: INFO nova.compute.manager [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Terminating instance [ 974.000065] env[62558]: DEBUG nova.compute.manager [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.000065] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.000065] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a336e5c9-3414-4712-b47a-b2922191c7c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.006658] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 974.006658] env[62558]: value = "task-1267144" [ 974.006658] env[62558]: _type = "Task" [ 974.006658] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.014824] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.060543] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 974.060543] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272598', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'name': 'volume-4d0add79-8ef3-4cf7-be38-7fd542cda129', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3f42034-3d2b-472a-89c0-5445cb6fb567', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'serial': '4d0add79-8ef3-4cf7-be38-7fd542cda129'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 974.060543] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc40a50-aac1-4fc1-b854-84bf3c06dfb1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.063171] env[62558]: DEBUG oslo_concurrency.lockutils [req-924d5366-f8a6-4de5-bc60-b80c65178b18 req-8aacfc8d-5fba-4bb7-a24f-4b0cc4f949da service nova] Releasing lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.063560] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.063731] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 974.081133] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de69e0c-b78a-4a4b-b510-dc46465e1414 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.108540] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] volume-4d0add79-8ef3-4cf7-be38-7fd542cda129/volume-4d0add79-8ef3-4cf7-be38-7fd542cda129.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.109246] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5763f9e4-b01e-45ed-90c1-588d339ce9cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.127990] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 974.127990] env[62558]: value = "task-1267145" [ 974.127990] env[62558]: _type = "Task" [ 974.127990] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.136152] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.251285] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.315801] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.317833] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.390s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.319406] env[62558]: INFO nova.compute.claims [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.337935] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033741} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.338927] env[62558]: INFO nova.scheduler.client.report [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted allocations for instance 0e5c1138-7f09-4b76-9c56-93c0ad947ad8 [ 974.339894] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.340085] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.340341] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk to [datastore1] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.342810] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e6655d6-9f61-4cb3-93c8-501dd6c7ea62 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.350428] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 974.350428] env[62558]: value = "task-1267146" [ 974.350428] env[62558]: _type = "Task" [ 974.350428] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.358770] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.516808] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267144, 'name': PowerOffVM_Task, 'duration_secs': 0.184818} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.516808] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.517097] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 974.517299] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272562', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'name': 'volume-5dc54ac6-cb87-4624-a362-73c54c37b21e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2480822a-e10a-4066-a5d8-5ca633ab9b12', 'attached_at': '', 'detached_at': '', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'serial': '5dc54ac6-cb87-4624-a362-73c54c37b21e'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 974.518053] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b34e188-16d3-4106-9a59-9a7d08eb5e3f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.535449] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5359d39-b115-49f4-8b1b-481b2272b1a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.541468] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b25d7e4-9f50-4f7f-a08f-4cf6a20c8922 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.559500] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2948fc9-0ff7-4934-99a1-ef2a6c37527d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.577511] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] The volume has not been displaced from its original location: [datastore2] volume-5dc54ac6-cb87-4624-a362-73c54c37b21e/volume-5dc54ac6-cb87-4624-a362-73c54c37b21e.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 974.582788] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Reconfiguring VM instance instance-00000059 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 974.583187] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-574bc216-2a8d-4d43-8f29-eb6676c781b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.601748] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 974.601748] env[62558]: value = "task-1267147" [ 974.601748] env[62558]: _type = "Task" [ 974.601748] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.613682] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.630231] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 974.642465] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267145, 'name': ReconfigVM_Task, 'duration_secs': 0.393351} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.642772] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfigured VM instance instance-00000056 to attach disk [datastore1] volume-4d0add79-8ef3-4cf7-be38-7fd542cda129/volume-4d0add79-8ef3-4cf7-be38-7fd542cda129.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.647966] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4641f2df-f860-4e97-9ef0-a4983d7efaab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.667856] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 974.667856] env[62558]: value = "task-1267148" [ 974.667856] env[62558]: _type = "Task" [ 974.667856] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.678214] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.848373] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7d629d5c-fdfd-4ace-8d53-afe5157e89db tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "0e5c1138-7f09-4b76-9c56-93c0ad947ad8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.527s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.864735] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.922647] env[62558]: DEBUG nova.network.neutron [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Updating instance_info_cache with network_info: [{"id": "a703fd04-54f4-4a70-a8d2-8b588f0b5658", "address": "fa:16:3e:72:23:6e", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa703fd04-54", "ovs_interfaceid": "a703fd04-54f4-4a70-a8d2-8b588f0b5658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.115864] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267147, 'name': ReconfigVM_Task, 'duration_secs': 0.174758} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.116197] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Reconfigured VM instance instance-00000059 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 975.121587] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-844d154e-3744-4510-9401-44893352f857 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.139718] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 975.139718] env[62558]: value = "task-1267149" [ 975.139718] env[62558]: _type = "Task" [ 975.139718] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.152347] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267149, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.178990] env[62558]: DEBUG oslo_vmware.api [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267148, 'name': ReconfigVM_Task, 'duration_secs': 0.151254} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.179389] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272598', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'name': 'volume-4d0add79-8ef3-4cf7-be38-7fd542cda129', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3f42034-3d2b-472a-89c0-5445cb6fb567', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'serial': '4d0add79-8ef3-4cf7-be38-7fd542cda129'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 975.365957] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.425772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "refresh_cache-ec2b48ff-dffa-4948-9925-a1c23aa76e64" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.426138] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance network_info: |[{"id": "a703fd04-54f4-4a70-a8d2-8b588f0b5658", "address": "fa:16:3e:72:23:6e", "network": {"id": "c4ddf8ed-a29d-482f-b2a0-60f09d171758", "bridge": "br-int", "label": "tempest-ImagesTestJSON-2147161641-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52e1d3e1dcd64e918c1e436e691e290b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa703fd04-54", "ovs_interfaceid": "a703fd04-54f4-4a70-a8d2-8b588f0b5658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 975.426595] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:23:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a703fd04-54f4-4a70-a8d2-8b588f0b5658', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.434538] env[62558]: DEBUG oslo.service.loopingcall [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.437652] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 975.438182] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8b6628c-5acf-421b-b2c7-a132e987c9a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.472441] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.472441] env[62558]: value = "task-1267150" [ 975.472441] env[62558]: _type = "Task" [ 975.472441] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.488396] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267150, 'name': CreateVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.635377] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0fdd7c-ea25-4e8b-992a-9ff1118071f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.647300] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44178a0d-8d12-4db7-bec4-c1c00dbcedec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.653610] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267149, 'name': ReconfigVM_Task, 'duration_secs': 0.178723} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.654881] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272562', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'name': 'volume-5dc54ac6-cb87-4624-a362-73c54c37b21e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2480822a-e10a-4066-a5d8-5ca633ab9b12', 'attached_at': '', 'detached_at': '', 'volume_id': '5dc54ac6-cb87-4624-a362-73c54c37b21e', 'serial': '5dc54ac6-cb87-4624-a362-73c54c37b21e'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 975.655215] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.656011] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53921f0f-ca8a-4be9-8016-fb57d40b0d5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.686787] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90fd673-545c-44a7-9375-02b78bcbfafc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.691521] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.692197] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a528b7da-a2ec-47cb-b21a-5fc844fe3694 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.699077] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccc4485-15e5-4926-b47f-11909f6a224b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.715572] env[62558]: DEBUG nova.compute.provider_tree [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.755061] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.755331] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.755513] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Deleting the datastore file [datastore2] 2480822a-e10a-4066-a5d8-5ca633ab9b12 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.755801] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1cf60361-2e4b-46b5-b96a-b09c9a76879b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.766348] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for the task: (returnval){ [ 975.766348] env[62558]: value = "task-1267152" [ 975.766348] env[62558]: _type = "Task" [ 975.766348] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.768042] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda60acf-2629-4ce8-b455-a42398d753b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.781409] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.798317] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.866433] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.985516] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267150, 'name': CreateVM_Task, 'duration_secs': 0.418546} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.985765] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.986578] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.986788] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.987238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.987471] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad3b37fc-ec2c-43e5-a5e1-36a1944cddc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.995112] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 975.995112] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f140b4-3456-cc5d-ab59-2be4f522160e" [ 975.995112] env[62558]: _type = "Task" [ 975.995112] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.006545] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f140b4-3456-cc5d-ab59-2be4f522160e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.033301] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.033577] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.033796] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.034038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.034229] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.036602] env[62558]: INFO nova.compute.manager [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Terminating instance [ 976.038593] env[62558]: DEBUG nova.compute.manager [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.038796] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.039658] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2574316-e067-4c9e-823d-7b5c6c443643 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.049618] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.049923] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d5d0c8f-ae15-48f1-8178-aba5685823b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.057757] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 976.057757] env[62558]: value = "task-1267153" [ 976.057757] env[62558]: _type = "Task" [ 976.057757] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.069367] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.218503] env[62558]: DEBUG nova.scheduler.client.report [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.226533] env[62558]: DEBUG nova.objects.instance [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid a3f42034-3d2b-472a-89c0-5445cb6fb567 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.278991] env[62558]: DEBUG oslo_vmware.api [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Task: {'id': task-1267152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152752} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.279330] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.279478] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.279663] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.279842] env[62558]: INFO nova.compute.manager [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Took 2.28 seconds to destroy the instance on the hypervisor. [ 976.280161] env[62558]: DEBUG oslo.service.loopingcall [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.280381] env[62558]: DEBUG nova.compute.manager [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.280479] env[62558]: DEBUG nova.network.neutron [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 976.304393] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.304758] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef4694c6-1928-407e-bb6b-d1480ad94069 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.313983] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 976.313983] env[62558]: value = "task-1267154" [ 976.313983] env[62558]: _type = "Task" [ 976.313983] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.324452] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.363404] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.509469] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f140b4-3456-cc5d-ab59-2be4f522160e, 'name': SearchDatastore_Task, 'duration_secs': 0.015125} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.509861] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.510188] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.510495] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.510704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.510940] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.511361] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b57a2229-e7f7-4bfb-be3c-badcd3e84004 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.524115] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.524115] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.524115] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06298c5b-70ab-4f5c-9c03-dc15ce34f20d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.532203] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 976.532203] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c77b1a-2057-001b-1dea-d7aa01f3d6e2" [ 976.532203] env[62558]: _type = "Task" [ 976.532203] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.543836] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c77b1a-2057-001b-1dea-d7aa01f3d6e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.569485] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267153, 'name': PowerOffVM_Task, 'duration_secs': 0.199614} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.569766] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.569953] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.570232] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2aeaa780-88a9-415b-bbf3-043c8a1cf435 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.628734] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.629061] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.629266] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleting the datastore file [datastore2] 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.629548] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93c5adbe-e410-4b7b-8368-51c8b7e44516 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.639813] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for the task: (returnval){ [ 976.639813] env[62558]: value = "task-1267156" [ 976.639813] env[62558]: _type = "Task" [ 976.639813] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.650616] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267156, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.723907] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.724630] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 976.728136] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.840s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.734666] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2f0e8508-920a-4d18-802e-40c52016b0c5 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.291s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.828425] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267154, 'name': PowerOffVM_Task, 'duration_secs': 0.174452} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.829069] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.829346] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 976.866764] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.900638] env[62558]: DEBUG nova.compute.manager [req-f4236cfd-402f-4f04-ac46-a295604e5cd9 req-c9e71a0c-0cd7-48f4-9331-6bc1627fa2ec service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Received event network-vif-deleted-1f94c71a-6387-41af-be7e-7fb674e8b3af {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.900849] env[62558]: INFO nova.compute.manager [req-f4236cfd-402f-4f04-ac46-a295604e5cd9 req-c9e71a0c-0cd7-48f4-9331-6bc1627fa2ec service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Neutron deleted interface 1f94c71a-6387-41af-be7e-7fb674e8b3af; detaching it from the instance and deleting it from the info cache [ 976.901041] env[62558]: DEBUG nova.network.neutron [req-f4236cfd-402f-4f04-ac46-a295604e5cd9 req-c9e71a0c-0cd7-48f4-9331-6bc1627fa2ec service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.043643] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c77b1a-2057-001b-1dea-d7aa01f3d6e2, 'name': SearchDatastore_Task, 'duration_secs': 0.01145} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.045669] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07556aae-ef73-46fb-b71c-129d2c5279f6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.050519] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 977.050519] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52498e38-ed70-e6d0-a9bd-94e34070772d" [ 977.050519] env[62558]: _type = "Task" [ 977.050519] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.058685] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52498e38-ed70-e6d0-a9bd-94e34070772d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.148623] env[62558]: DEBUG oslo_vmware.api [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Task: {'id': task-1267156, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171928} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.151800] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.151800] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 977.151800] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 977.151800] env[62558]: INFO nova.compute.manager [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 977.151800] env[62558]: DEBUG oslo.service.loopingcall [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.151800] env[62558]: DEBUG nova.compute.manager [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.151800] env[62558]: DEBUG nova.network.neutron [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 977.234020] env[62558]: DEBUG nova.compute.utils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.234020] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.234020] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 977.281090] env[62558]: DEBUG nova.policy [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3faaab5e759446789a31b57b0abe52e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebd5df2c8e5f42f5b4b3da70f8793ed9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.338242] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.338942] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.339296] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.341901] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.345157] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a6ed1c2-6794-403d-a975-7b21067f6ced {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.369409] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267146, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.604201} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.370883] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bf57715a-d9cb-4e6c-885f-8cbb16b14efd/bf57715a-d9cb-4e6c-885f-8cbb16b14efd.vmdk to [datastore1] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.371480] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 977.371480] env[62558]: value = "task-1267157" [ 977.371480] env[62558]: _type = "Task" [ 977.371480] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.372588] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e30b163-29eb-4f98-8a6c-e2b0886f4a97 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.379962] env[62558]: DEBUG nova.network.neutron [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.405102] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.413023] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0760869-4d41-4436-8897-1e573d902c85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.427444] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267157, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.427906] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-09e1ab41-5781-4bd2-a884-aa8c24059a52 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.438946] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521ffe32-375e-4344-ab7f-d798a60d3555 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.450542] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 977.450542] env[62558]: value = "task-1267158" [ 977.450542] env[62558]: _type = "Task" [ 977.450542] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.459788] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267158, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.472504] env[62558]: DEBUG nova.compute.manager [req-f4236cfd-402f-4f04-ac46-a295604e5cd9 req-c9e71a0c-0cd7-48f4-9331-6bc1627fa2ec service nova] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Detach interface failed, port_id=1f94c71a-6387-41af-be7e-7fb674e8b3af, reason: Instance 2480822a-e10a-4066-a5d8-5ca633ab9b12 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 977.560184] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52498e38-ed70-e6d0-a9bd-94e34070772d, 'name': SearchDatastore_Task, 'duration_secs': 0.015333} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.560473] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.562320] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ec2b48ff-dffa-4948-9925-a1c23aa76e64/ec2b48ff-dffa-4948-9925-a1c23aa76e64.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.562320] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f5c8a81-0a09-48c6-88eb-76d75f730487 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.568307] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 977.568307] env[62558]: value = "task-1267159" [ 977.568307] env[62558]: _type = "Task" [ 977.568307] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.576853] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267159, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.597843] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Successfully created port: c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.744294] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.748968] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Applying migration context for instance c65dbe09-d073-48ad-a18a-8b6383c7e345 as it has an incoming, in-progress migration 9b9fa3c1-4591-4035-8fc3-73041d59ebdd. Migration status is migrating {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 977.750893] env[62558]: INFO nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating resource usage from migration 9b9fa3c1-4591-4035-8fc3-73041d59ebdd [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ac801d7-af70-46e0-88b3-02caee13497d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4815ba3f-265f-466a-9850-4c325cdb88de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a3f42034-3d2b-472a-89c0-5445cb6fb567 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2480822a-e10a-4066-a5d8-5ca633ab9b12 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 29584082-37e6-4dd8-906f-fd87b4ca9bc2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 109f67db-21cf-40f9-966e-4730bff98bcf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 591d326c-c20c-401b-a57b-895d59c3b418 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 49a58b46-207f-4515-b313-afcdb2d1ced3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789019] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ef0b3d16-704d-4435-9c23-a258d94a9983 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 977.789019] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ec2b48ff-dffa-4948-9925-a1c23aa76e64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789457] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Migration 9b9fa3c1-4591-4035-8fc3-73041d59ebdd is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 977.789457] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance c65dbe09-d073-48ad-a18a-8b6383c7e345 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789457] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e8b3008-693d-4445-937b-8e7db13c7c91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 977.789712] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 977.790011] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3456MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 977.883573] env[62558]: INFO nova.compute.manager [-] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Took 1.60 seconds to deallocate network for instance. [ 977.897260] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267157, 'name': ReconfigVM_Task, 'duration_secs': 0.417751} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.897260] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 977.969478] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267158, 'name': ReconfigVM_Task, 'duration_secs': 0.493241} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.973529] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 49a58b46-207f-4515-b313-afcdb2d1ced3/49a58b46-207f-4515-b313-afcdb2d1ced3.vmdk or device None with type streamOptimized {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.975165] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5d40d19-8710-42f0-9f7b-e7faacda3c63 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.985430] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 977.985430] env[62558]: value = "task-1267160" [ 977.985430] env[62558]: _type = "Task" [ 977.985430] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.002562] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267160, 'name': Rename_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.080860] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267159, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466434} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.080860] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ec2b48ff-dffa-4948-9925-a1c23aa76e64/ec2b48ff-dffa-4948-9925-a1c23aa76e64.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.081020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.081306] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc088657-521f-4763-a710-6dd6cb253023 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.088325] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 978.088325] env[62558]: value = "task-1267161" [ 978.088325] env[62558]: _type = "Task" [ 978.088325] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.096847] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.099472] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba48af27-a2eb-47b9-b4a3-a2b2f3312d7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.106490] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec72bcad-5fc8-4cea-b8aa-e611529b272d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.136770] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aedb6d2-0a2a-4ca3-8176-630efc9367b1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.145018] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05575ea-83b4-42d6-b4d0-d222d5806668 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.160280] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.349998] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "ea680004-b94a-47f9-b516-2afb95c30a51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.350299] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.410364] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.410364] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.410364] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.410364] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.410944] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.411286] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.412490] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.412490] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.412490] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.412490] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.412490] env[62558]: DEBUG nova.virt.hardware [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.418130] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 978.418872] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a15a2c5b-1cbf-4f1a-b8f9-7f4357290024 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.446117] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 978.446117] env[62558]: value = "task-1267162" [ 978.446117] env[62558]: _type = "Task" [ 978.446117] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.447320] env[62558]: INFO nova.compute.manager [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Took 0.56 seconds to detach 1 volumes for instance. [ 978.457020] env[62558]: DEBUG nova.compute.manager [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Deleting volume: 5dc54ac6-cb87-4624-a362-73c54c37b21e {{(pid=62558) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 978.466529] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267162, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.501622] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267160, 'name': Rename_Task, 'duration_secs': 0.193453} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.501945] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.502228] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8754964-a22a-45a9-9765-c5ad3b8c55e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.509257] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 978.509257] env[62558]: value = "task-1267163" [ 978.509257] env[62558]: _type = "Task" [ 978.509257] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.517287] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.558944] env[62558]: DEBUG nova.network.neutron [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.598081] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063135} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.598081] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.598713] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c5e095-1cf4-48d2-9dd8-6be2621fa337 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.621811] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] ec2b48ff-dffa-4948-9925-a1c23aa76e64/ec2b48ff-dffa-4948-9925-a1c23aa76e64.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.623035] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9e2991c-2eea-459f-a47b-1789692de466 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.641028] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 978.641028] env[62558]: value = "task-1267165" [ 978.641028] env[62558]: _type = "Task" [ 978.641028] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.649257] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.663450] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.765808] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 978.790479] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.790479] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.790635] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.790741] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.790897] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.791058] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.791271] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.791433] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.791602] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.791771] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.792015] env[62558]: DEBUG nova.virt.hardware [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.792878] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73d7b68-b0c8-4cf7-b7a6-f7eb1f1c837f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.801017] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872ce719-bd66-4b9d-a3e1-a2950c4458cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.853230] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.928835] env[62558]: DEBUG nova.compute.manager [req-84d7b6f3-fcd8-4d75-a0c8-354afdef392a req-3efdb9e7-410a-418e-aeb3-39bf4b71aa3e service nova] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Received event network-vif-deleted-7472a75c-48b2-4f52-a2b5-905ae7a9ef3b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.958845] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267162, 'name': ReconfigVM_Task, 'duration_secs': 0.160034} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.959169] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 978.959927] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73175850-d11f-4a13-9f40-2a63c4355065 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.983714] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.984043] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b752c3f0-45a2-4c7a-9610-7a28c5ab4112 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.000492] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.006585] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 979.006585] env[62558]: value = "task-1267166" [ 979.006585] env[62558]: _type = "Task" [ 979.006585] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.017726] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.020728] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267163, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.062401] env[62558]: INFO nova.compute.manager [-] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Took 1.91 seconds to deallocate network for instance. [ 979.151028] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267165, 'name': ReconfigVM_Task, 'duration_secs': 0.289142} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.151479] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Reconfigured VM instance instance-00000061 to attach disk [datastore2] ec2b48ff-dffa-4948-9925-a1c23aa76e64/ec2b48ff-dffa-4948-9925-a1c23aa76e64.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.153451] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b27dd1d-c1e9-4250-a29f-4dcde1dac56b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.161710] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 979.161710] env[62558]: value = "task-1267167" [ 979.161710] env[62558]: _type = "Task" [ 979.161710] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.171199] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 979.171633] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.443s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.171958] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267167, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.172376] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.287s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.173139] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.174664] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.280s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.174918] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.176894] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.177s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.177211] env[62558]: DEBUG nova.objects.instance [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lazy-loading 'resources' on Instance uuid 2480822a-e10a-4066-a5d8-5ca633ab9b12 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.185285] env[62558]: DEBUG nova.compute.manager [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Received event network-vif-plugged-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.185512] env[62558]: DEBUG oslo_concurrency.lockutils [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.185725] env[62558]: DEBUG oslo_concurrency.lockutils [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.185894] env[62558]: DEBUG oslo_concurrency.lockutils [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.186341] env[62558]: DEBUG nova.compute.manager [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] No waiting events found dispatching network-vif-plugged-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 979.187287] env[62558]: WARNING nova.compute.manager [req-910dd682-ceb6-418d-8f5a-918b819ea6c4 req-489bc332-f3a0-4343-8489-fe9f1e97cb4a service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Received unexpected event network-vif-plugged-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 for instance with vm_state building and task_state spawning. [ 979.190554] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.193062] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Successfully updated port: c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.199353] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.199615] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.199947] env[62558]: DEBUG nova.objects.instance [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.218273] env[62558]: INFO nova.scheduler.client.report [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance 591d326c-c20c-401b-a57b-895d59c3b418 [ 979.220022] env[62558]: INFO nova.scheduler.client.report [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance ef0b3d16-704d-4435-9c23-a258d94a9983 [ 979.373990] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.519780] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267166, 'name': ReconfigVM_Task, 'duration_secs': 0.281925} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.523077] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Reconfigured VM instance instance-0000005e to attach disk [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345/c65dbe09-d073-48ad-a18a-8b6383c7e345.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.523077] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 979.526442] env[62558]: DEBUG oslo_vmware.api [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267163, 'name': PowerOnVM_Task, 'duration_secs': 0.687087} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.526866] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.572264] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.621921] env[62558]: DEBUG nova.compute.manager [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.622950] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7304f592-bb6e-430d-be1d-6e6129b78b79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.672093] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267167, 'name': Rename_Task, 'duration_secs': 0.226203} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.672263] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.672520] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fad5e4f8-8dcd-48c3-abec-6f1e8e529e36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.679626] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 979.679626] env[62558]: value = "task-1267168" [ 979.679626] env[62558]: _type = "Task" [ 979.679626] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.690566] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.698270] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.698660] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.698660] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.705499] env[62558]: DEBUG nova.objects.instance [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.733179] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1e956c33-9858-4338-82f2-0b618d9d2f6c tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "591d326c-c20c-401b-a57b-895d59c3b418" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.247s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.734578] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aef63da3-db1c-414f-a62e-686f156acb29 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ef0b3d16-704d-4435-9c23-a258d94a9983" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.741s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.887989] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ad3a0-6967-4519-b931-c0fb9b30fcf3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.896662] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292d5f9e-18d0-4587-a159-6b0a3df0bcea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.926101] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9016895b-ef16-45e7-9d43-c117fb868fe9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.933986] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3bce3a-3c85-43d7-b742-79055084eee4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.950404] env[62558]: DEBUG nova.compute.provider_tree [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.030798] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad96d53e-8a61-4ca4-8932-baf4d164e838 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.051918] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18862558-c735-4069-88e7-e806c7845267 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.070615] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 980.143784] env[62558]: DEBUG oslo_concurrency.lockutils [None req-67d4d56a-ec4e-492b-b3ec-648569f71ca9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.300s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.189884] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267168, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.210088] env[62558]: DEBUG nova.objects.base [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance<29584082-37e6-4dd8-906f-fd87b4ca9bc2> lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 980.210321] env[62558]: DEBUG nova.network.neutron [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 980.237932] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.333772] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f8c16fc2-280c-47db-8a61-96be5b6633c8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.134s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.375442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.375442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.378906] env[62558]: DEBUG nova.network.neutron [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating instance_info_cache with network_info: [{"id": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "address": "fa:16:3e:25:4b:2d", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e25a3a-b0", "ovs_interfaceid": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.456744] env[62558]: DEBUG nova.scheduler.client.report [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.610178] env[62558]: DEBUG nova.network.neutron [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Port 46f92558-3c85-4803-aad8-7ba3157e3f67 binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 980.632378] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.632676] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.632851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.633043] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.633219] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.635670] env[62558]: INFO nova.compute.manager [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Terminating instance [ 980.637856] env[62558]: DEBUG nova.compute.manager [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.638073] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.638942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4b8d60-714f-475a-83b8-81eaf5282823 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.648157] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.648377] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9d92e9a-73fd-439f-ac53-45c86bc199b3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.655462] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 980.655462] env[62558]: value = "task-1267169" [ 980.655462] env[62558]: _type = "Task" [ 980.655462] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.365457] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.368221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.368597] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Instance network_info: |[{"id": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "address": "fa:16:3e:25:4b:2d", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e25a3a-b0", "ovs_interfaceid": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.370377] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.193s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.375874] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.376092] env[62558]: WARNING oslo_vmware.common.loopingcall [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] task run outlasted interval by 0.22038899999999995 sec [ 981.381032] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:4b:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53915f38-d7a0-42ec-8b30-1eacfb2cc379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8e25a3a-b0f5-428b-a485-ff34df8b7fe5', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.388747] env[62558]: DEBUG oslo.service.loopingcall [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.391428] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.018s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.393091] env[62558]: INFO nova.compute.claims [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.397709] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 981.401569] env[62558]: INFO nova.scheduler.client.report [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Deleted allocations for instance 2480822a-e10a-4066-a5d8-5ca633ab9b12 [ 981.405837] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20bbd966-d66e-48b4-acc4-d92e666dfd7c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.426649] env[62558]: DEBUG oslo_vmware.api [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267168, 'name': PowerOnVM_Task, 'duration_secs': 0.595536} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.427378] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.427631] env[62558]: INFO nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Took 9.37 seconds to spawn the instance on the hypervisor. [ 981.427881] env[62558]: DEBUG nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.431329] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1152a157-b9e2-416e-8b19-c808f47bf6eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.434074] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267169, 'name': PowerOffVM_Task, 'duration_secs': 0.210345} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.435268] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.435461] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.435936] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.435936] env[62558]: value = "task-1267170" [ 981.435936] env[62558]: _type = "Task" [ 981.435936] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.436513] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-620d1426-a894-4715-859a-03928aa74987 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.453383] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267170, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.473353] env[62558]: DEBUG nova.compute.manager [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Received event network-changed-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.473533] env[62558]: DEBUG nova.compute.manager [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Refreshing instance network info cache due to event network-changed-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.473781] env[62558]: DEBUG oslo_concurrency.lockutils [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] Acquiring lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.474014] env[62558]: DEBUG oslo_concurrency.lockutils [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] Acquired lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.474341] env[62558]: DEBUG nova.network.neutron [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Refreshing network info cache for port c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 981.518866] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.519026] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.519324] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore1] ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.520065] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-315e08d3-9aea-4eae-a4a6-967d21645ae2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.526685] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 981.526685] env[62558]: value = "task-1267172" [ 981.526685] env[62558]: _type = "Task" [ 981.526685] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.535266] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.885576] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.892189] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.892433] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.909262] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.909497] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.909675] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.919564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2c11cf9e-1dcd-4211-b1e4-959bf8ef4eea tempest-ServersTestBootFromVolume-1669886052 tempest-ServersTestBootFromVolume-1669886052-project-member] Lock "2480822a-e10a-4066-a5d8-5ca633ab9b12" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.928s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.947177] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267170, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.954191] env[62558]: INFO nova.compute.manager [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Took 22.96 seconds to build instance. [ 982.036541] env[62558]: DEBUG oslo_vmware.api [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145072} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.036807] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.037386] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.038031] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.038031] env[62558]: INFO nova.compute.manager [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Took 1.40 seconds to destroy the instance on the hypervisor. [ 982.038161] env[62558]: DEBUG oslo.service.loopingcall [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.038389] env[62558]: DEBUG nova.compute.manager [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.038389] env[62558]: DEBUG nova.network.neutron [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.209389] env[62558]: DEBUG nova.network.neutron [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updated VIF entry in instance network info cache for port c8e25a3a-b0f5-428b-a485-ff34df8b7fe5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.209389] env[62558]: DEBUG nova.network.neutron [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating instance_info_cache with network_info: [{"id": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "address": "fa:16:3e:25:4b:2d", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e25a3a-b0", "ovs_interfaceid": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.395146] env[62558]: INFO nova.compute.manager [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Detaching volume c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78 [ 982.432648] env[62558]: INFO nova.virt.block_device [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Attempting to driver detach volume c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78 from mountpoint /dev/sdb [ 982.432945] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 982.433198] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272572', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'name': 'volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4815ba3f-265f-466a-9850-4c325cdb88de', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'serial': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 982.434401] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de411196-2be1-4a24-be3c-7456597c9187 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.465869] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5e5774c6-1abd-4983-9eaa-be5e09a8e630 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.482s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.470559] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e02334b-2efd-49d6-a4de-8bf250182b43 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.477452] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267170, 'name': CreateVM_Task, 'duration_secs': 0.72201} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.477956] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.478652] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.478823] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.479312] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.479606] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c65abffa-a6a4-4eeb-a705-316edfaa0566 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.485899] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a3b8fb-9791-41c3-8ad5-923dede5edb9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.493409] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 982.493409] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e36f51-3b33-b146-8119-fc0d05efecdd" [ 982.493409] env[62558]: _type = "Task" [ 982.493409] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.517634] env[62558]: DEBUG nova.compute.manager [req-19f5db37-7873-449f-ba4f-db4ad62d34bc req-46305664-f010-4eb2-8eb2-fe405f3522af service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Received event network-vif-deleted-2fc3029c-f3d0-4f21-91c0-1328caff5374 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.517883] env[62558]: INFO nova.compute.manager [req-19f5db37-7873-449f-ba4f-db4ad62d34bc req-46305664-f010-4eb2-8eb2-fe405f3522af service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Neutron deleted interface 2fc3029c-f3d0-4f21-91c0-1328caff5374; detaching it from the instance and deleting it from the info cache [ 982.518106] env[62558]: DEBUG nova.network.neutron [req-19f5db37-7873-449f-ba4f-db4ad62d34bc req-46305664-f010-4eb2-8eb2-fe405f3522af service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.519653] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d7c8f8-429f-45d5-b9cc-56cb16cf7f6d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.526659] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e36f51-3b33-b146-8119-fc0d05efecdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.540269] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] The volume has not been displaced from its original location: [datastore2] volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78/volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 982.545782] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 982.549222] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e00e847b-0d20-4d79-8954-e42b246fd724 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.566616] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9ceff7-4d33-4b42-a119-989aa151ec20 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.572886] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 982.572886] env[62558]: value = "task-1267173" [ 982.572886] env[62558]: _type = "Task" [ 982.572886] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.576989] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Suspending the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 982.580681] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ae32920b-2f64-4a34-9085-fc5b85b80c7d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.592436] env[62558]: DEBUG oslo_vmware.api [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 982.592436] env[62558]: value = "task-1267174" [ 982.592436] env[62558]: _type = "Task" [ 982.592436] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.592630] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267173, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.606171] env[62558]: DEBUG oslo_vmware.api [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267174, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.710950] env[62558]: DEBUG oslo_concurrency.lockutils [req-8aad84f6-0b41-411c-b5e5-098524022a58 req-b7ed4792-093c-49cb-9401-4c72b1b7e6e7 service nova] Releasing lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.736881] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017acc4c-07b8-41ff-8dbe-956957340973 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.745557] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f969289-5e36-4762-8ea7-da5ee0c15284 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.749574] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.749832] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.750178] env[62558]: DEBUG nova.objects.instance [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.788241] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dbc14a-0210-4a3c-a664-4e9dbc5c49bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.796743] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb2ab58-2a10-4adc-909b-7d81d89ddb1e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.812549] env[62558]: DEBUG nova.compute.provider_tree [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.881949] env[62558]: DEBUG nova.network.neutron [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.960024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.960024] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.960024] env[62558]: DEBUG nova.network.neutron [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.006797] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52e36f51-3b33-b146-8119-fc0d05efecdd, 'name': SearchDatastore_Task, 'duration_secs': 0.038265} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.007149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.007413] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.007693] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.007862] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.008089] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.008435] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63992661-77be-4a99-b689-cee5a832b5ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.018851] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.019134] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.019975] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cb346ac-0a02-435d-a3e1-cd47009c8d8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.027735] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9248ba7-53d8-49dc-8d18-990f419fece1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.029790] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 983.029790] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5252bcff-148c-0a11-e546-858fc643d615" [ 983.029790] env[62558]: _type = "Task" [ 983.029790] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.037744] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e9972a-b5b7-4bfa-acc2-df9e18e29703 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.052583] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5252bcff-148c-0a11-e546-858fc643d615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.076283] env[62558]: DEBUG nova.compute.manager [req-19f5db37-7873-449f-ba4f-db4ad62d34bc req-46305664-f010-4eb2-8eb2-fe405f3522af service nova] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Detach interface failed, port_id=2fc3029c-f3d0-4f21-91c0-1328caff5374, reason: Instance ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 983.086547] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267173, 'name': ReconfigVM_Task, 'duration_secs': 0.275158} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.086867] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 983.091990] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afe5c1f6-18a1-47bb-89fe-c9de6cc5de69 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.113926] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 983.113926] env[62558]: value = "task-1267175" [ 983.113926] env[62558]: _type = "Task" [ 983.113926] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.117781] env[62558]: DEBUG oslo_vmware.api [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267174, 'name': SuspendVM_Task} progress is 62%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.128860] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267175, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.316282] env[62558]: DEBUG nova.scheduler.client.report [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.361853] env[62558]: DEBUG nova.objects.instance [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.385739] env[62558]: INFO nova.compute.manager [-] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Took 1.35 seconds to deallocate network for instance. [ 983.541736] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5252bcff-148c-0a11-e546-858fc643d615, 'name': SearchDatastore_Task, 'duration_secs': 0.009924} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.542698] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83f6e846-cf2a-474d-927b-66f70d05ee1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.548839] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 983.548839] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529daa70-3466-135b-2aea-b8eb761cd78a" [ 983.548839] env[62558]: _type = "Task" [ 983.548839] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.558454] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529daa70-3466-135b-2aea-b8eb761cd78a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.613166] env[62558]: DEBUG oslo_vmware.api [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267174, 'name': SuspendVM_Task, 'duration_secs': 0.698999} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.613649] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Suspended the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 983.613833] env[62558]: DEBUG nova.compute.manager [None req-33b10596-2f22-4360-bac0-38d31c838b48 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.616644] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e30569-5c1b-4288-9bf3-31ff5e1ba924 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.631934] env[62558]: DEBUG oslo_vmware.api [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267175, 'name': ReconfigVM_Task, 'duration_secs': 0.16056} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.635586] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272572', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'name': 'volume-c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4815ba3f-265f-466a-9850-4c325cdb88de', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78', 'serial': 'c8fb33e4-f9bb-4378-a6ac-9a967bbf1e78'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 983.778909] env[62558]: DEBUG nova.network.neutron [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.821643] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.823482] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.826209] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.254s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.827300] env[62558]: DEBUG nova.objects.instance [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lazy-loading 'resources' on Instance uuid 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.861824] env[62558]: DEBUG nova.objects.base [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance<29584082-37e6-4dd8-906f-fd87b4ca9bc2> lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 983.861824] env[62558]: DEBUG nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.894870] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.062503] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]529daa70-3466-135b-2aea-b8eb761cd78a, 'name': SearchDatastore_Task, 'duration_secs': 0.049482} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.062764] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.063029] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e8b3008-693d-4445-937b-8e7db13c7c91/4e8b3008-693d-4445-937b-8e7db13c7c91.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.063309] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88678f7f-cb97-48bc-adfc-ad2ec2b67946 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.073178] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 984.073178] env[62558]: value = "task-1267176" [ 984.073178] env[62558]: _type = "Task" [ 984.073178] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.087350] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.095561] env[62558]: DEBUG nova.policy [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.115390] env[62558]: DEBUG nova.compute.manager [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.116714] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57304fa-abf3-49bb-9c40-0e2fd00b3bba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.191280] env[62558]: DEBUG nova.objects.instance [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'flavor' on Instance uuid 4815ba3f-265f-466a-9850-4c325cdb88de {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.281902] env[62558]: DEBUG oslo_concurrency.lockutils [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.330862] env[62558]: DEBUG nova.compute.utils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.335593] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 984.335758] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.378134] env[62558]: DEBUG nova.policy [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37bb028d611e42d98319c4afba5b1b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46aa5789fccf4925aa49f9c4050c8463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.413793] env[62558]: DEBUG nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Successfully created port: 5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.536707] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71759675-7ef6-43a4-9334-265d7e1e75b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.544591] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83e5835-eb57-449e-aaf4-132498b36506 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.577214] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ed24f9-44a6-4064-a794-d1673b688c70 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.585520] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.588433] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58128ad8-6f29-49a1-8e2d-3025f046e5c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.603577] env[62558]: DEBUG nova.compute.provider_tree [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.631864] env[62558]: INFO nova.compute.manager [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] instance snapshotting [ 984.637157] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcdea1b-be1d-4855-abd4-756ae1fa34b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.659992] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5240496-3315-411e-ba04-eaf1b410c869 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.664470] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Successfully created port: b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.804351] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a137f886-32e1-4356-bb60-46318dd7af61 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.834897] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6103fc43-24bc-4f45-93fd-6dfba4a43c5d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.838854] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.847786] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 985.085966] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267176, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.107246] env[62558]: DEBUG nova.scheduler.client.report [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.173338] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Creating Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 985.173682] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b7f159f6-3c9e-49d1-80ac-4cfdb7fd0aee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.182305] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 985.182305] env[62558]: value = "task-1267177" [ 985.182305] env[62558]: _type = "Task" [ 985.182305] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.191929] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267177, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.203548] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3fee0fca-8f6b-4915-8ca5-c5f89fbbaca1 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.311s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.355372] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 985.355718] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11232ce8-67bc-479b-a021-2df7dcfa45ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.367545] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 985.367545] env[62558]: value = "task-1267178" [ 985.367545] env[62558]: _type = "Task" [ 985.367545] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.377241] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.500667] env[62558]: INFO nova.compute.manager [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Resuming [ 985.500953] env[62558]: DEBUG nova.objects.instance [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'flavor' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.584731] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267176, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.021806} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.584731] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e8b3008-693d-4445-937b-8e7db13c7c91/4e8b3008-693d-4445-937b-8e7db13c7c91.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.584731] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.584988] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75800a57-0f63-4a04-ab50-239320d07a54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.592693] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 985.592693] env[62558]: value = "task-1267179" [ 985.592693] env[62558]: _type = "Task" [ 985.592693] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.601525] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267179, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.612623] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.615582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.730s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.617406] env[62558]: INFO nova.compute.claims [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.637048] env[62558]: INFO nova.scheduler.client.report [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Deleted allocations for instance 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5 [ 985.694807] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267177, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.853333] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.880208] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267178, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.890146] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.890146] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.890323] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.890885] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.890885] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.890885] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.891019] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.891154] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.893682] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.893682] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.893682] env[62558]: DEBUG nova.virt.hardware [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.893682] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fed900-2868-48e3-8d30-d305954369c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.901697] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b30eab-9ef6-4b43-b666-104d302e191b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.998650] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.998941] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.999174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.999390] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.999549] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.002067] env[62558]: INFO nova.compute.manager [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Terminating instance [ 986.004619] env[62558]: DEBUG nova.compute.manager [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.004853] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.012888] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502e22f2-794e-4237-ba25-69617e810079 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.024619] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.024898] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1de06526-9185-4785-a087-850ed8a13ab5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.032943] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 986.032943] env[62558]: value = "task-1267180" [ 986.032943] env[62558]: _type = "Task" [ 986.032943] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.043036] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.102876] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267179, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067367} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.103690] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.104071] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe149740-f6e6-4887-be8a-b034f17c89b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.127066] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 4e8b3008-693d-4445-937b-8e7db13c7c91/4e8b3008-693d-4445-937b-8e7db13c7c91.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.128096] env[62558]: DEBUG nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Successfully updated port: 5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.131504] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-317fb620-3155-4138-be62-dd98d36b1ae0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.153472] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6aab40d3-5ac4-4ac0-89f3-229c71857273 tempest-AttachVolumeShelveTestJSON-1338984042 tempest-AttachVolumeShelveTestJSON-1338984042-project-member] Lock "8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.120s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.161221] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 986.161221] env[62558]: value = "task-1267181" [ 986.161221] env[62558]: _type = "Task" [ 986.161221] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.173680] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267181, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.177499] env[62558]: DEBUG nova.compute.manager [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-plugged-5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.177757] env[62558]: DEBUG oslo_concurrency.lockutils [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.177989] env[62558]: DEBUG oslo_concurrency.lockutils [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.178345] env[62558]: DEBUG oslo_concurrency.lockutils [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.178578] env[62558]: DEBUG nova.compute.manager [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] No waiting events found dispatching network-vif-plugged-5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.178843] env[62558]: WARNING nova.compute.manager [req-595840ca-abdf-4ae0-9095-f40b72e9643b req-488381c9-445c-4134-bb5d-96f20e160fbe service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received unexpected event network-vif-plugged-5f5f33c6-f6bc-4194-9da2-9e3bca010a13 for instance with vm_state active and task_state None. [ 986.195287] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267177, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.217483] env[62558]: DEBUG nova.compute.manager [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Received event network-vif-plugged-b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.217713] env[62558]: DEBUG oslo_concurrency.lockutils [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] Acquiring lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.217927] env[62558]: DEBUG oslo_concurrency.lockutils [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] Lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.218463] env[62558]: DEBUG oslo_concurrency.lockutils [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] Lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.218700] env[62558]: DEBUG nova.compute.manager [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] No waiting events found dispatching network-vif-plugged-b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.218875] env[62558]: WARNING nova.compute.manager [req-62948669-72b6-4d8d-81a8-772be9bf1898 req-d1a18b28-f201-4b18-b3c2-06fae0d6d0e1 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Received unexpected event network-vif-plugged-b2ae5c23-49aa-4034-b884-b1017e034d61 for instance with vm_state building and task_state spawning. [ 986.241275] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Successfully updated port: b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.362500] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7def56d-2aeb-4b11-83db-6d353faaa1a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.375403] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b62bb6-1547-4f98-b576-46a175a4669a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.387963] env[62558]: DEBUG oslo_vmware.api [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267178, 'name': PowerOnVM_Task, 'duration_secs': 0.581847} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.416204] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.417081] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-321893e9-b1d6-4c1f-b5c9-7e6fa77c2484 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance 'c65dbe09-d073-48ad-a18a-8b6383c7e345' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 986.421225] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388d8745-a37c-4415-aff1-4003e433f6dd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.430079] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be4e586-a06f-4e59-96a8-a326fd921fda {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.445113] env[62558]: DEBUG nova.compute.provider_tree [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.518495] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.518705] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquired lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.518882] env[62558]: DEBUG nova.network.neutron [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.542864] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267180, 'name': PowerOffVM_Task, 'duration_secs': 0.504911} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.543173] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.543354] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.543647] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aad26b0a-7120-4dab-a37b-2118c5e4bdab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.609228] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.609477] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.609664] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleting the datastore file [datastore2] 4815ba3f-265f-466a-9850-4c325cdb88de {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.609946] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03a55314-8824-4d82-97c4-ed3f7630bb6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.617737] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 986.617737] env[62558]: value = "task-1267183" [ 986.617737] env[62558]: _type = "Task" [ 986.617737] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.626777] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.647680] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.647892] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.648095] env[62558]: DEBUG nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.677102] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267181, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.694820] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267177, 'name': CreateSnapshot_Task, 'duration_secs': 1.058447} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.695621] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Created Snapshot of the VM instance {{(pid=62558) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 986.696579] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3837f168-ccf5-4a19-ad8d-db4460148313 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.746773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.746920] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.747087] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.948372] env[62558]: DEBUG nova.scheduler.client.report [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.129761] env[62558]: DEBUG oslo_vmware.api [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1958} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.130064] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.130221] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.130399] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.130573] env[62558]: INFO nova.compute.manager [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Took 1.13 seconds to destroy the instance on the hypervisor. [ 987.130816] env[62558]: DEBUG oslo.service.loopingcall [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.131014] env[62558]: DEBUG nova.compute.manager [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.131113] env[62558]: DEBUG nova.network.neutron [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.177638] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267181, 'name': ReconfigVM_Task, 'duration_secs': 0.640819} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.177927] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 4e8b3008-693d-4445-937b-8e7db13c7c91/4e8b3008-693d-4445-937b-8e7db13c7c91.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.178605] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a01e5fb-0b58-4d39-8953-dd8bf356bc3f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.185667] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 987.185667] env[62558]: value = "task-1267184" [ 987.185667] env[62558]: _type = "Task" [ 987.185667] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.194442] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267184, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.202287] env[62558]: WARNING nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 987.215080] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Creating linked-clone VM from snapshot {{(pid=62558) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 987.215435] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d61494d0-5293-4c24-bfbc-2d6a79d44585 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.225329] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 987.225329] env[62558]: value = "task-1267185" [ 987.225329] env[62558]: _type = "Task" [ 987.225329] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.236534] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267185, 'name': CloneVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.241050] env[62558]: DEBUG nova.network.neutron [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [{"id": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "address": "fa:16:3e:85:85:e2", "network": {"id": "3de64f77-2172-44c8-8ec7-c1d07cc22ca7", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1566618797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7663e884862d43b49eabe6396d8cd107", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53ebf5df-5ecb-4a0c-a163-d88165639de0", "external-id": "nsx-vlan-transportzone-588", "segmentation_id": 588, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37703c4e-05", "ovs_interfaceid": "37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.284331] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.430961] env[62558]: DEBUG nova.network.neutron [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updating instance_info_cache with network_info: [{"id": "b2ae5c23-49aa-4034-b884-b1017e034d61", "address": "fa:16:3e:92:d4:32", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ae5c23-49", "ovs_interfaceid": "b2ae5c23-49aa-4034-b884-b1017e034d61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.453344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.838s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.455623] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 987.457310] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.563s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.457745] env[62558]: DEBUG nova.objects.instance [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'resources' on Instance uuid ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.697487] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267184, 'name': Rename_Task, 'duration_secs': 0.380879} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.698237] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.700368] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-623c0a5d-4309-451e-a3b6-7df915b17e9d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.713380] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 987.713380] env[62558]: value = "task-1267186" [ 987.713380] env[62558]: _type = "Task" [ 987.713380] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.726286] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.734992] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267185, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.743792] env[62558]: DEBUG oslo_concurrency.lockutils [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Releasing lock "refresh_cache-49a58b46-207f-4515-b313-afcdb2d1ced3" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.744812] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b0661b-bb29-4f83-9693-49bb7da6cdd4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.752475] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Resuming the VM {{(pid=62558) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 987.752908] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b512fef-7f84-4f4f-a7a0-58bcff649e12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.761535] env[62558]: DEBUG oslo_vmware.api [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 987.761535] env[62558]: value = "task-1267187" [ 987.761535] env[62558]: _type = "Task" [ 987.761535] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.774973] env[62558]: DEBUG oslo_vmware.api [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267187, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.931289] env[62558]: DEBUG nova.network.neutron [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "address": "fa:16:3e:e0:7b:e2", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f5f33c6-f6", "ovs_interfaceid": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.940428] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.940902] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Instance network_info: |[{"id": "b2ae5c23-49aa-4034-b884-b1017e034d61", "address": "fa:16:3e:92:d4:32", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ae5c23-49", "ovs_interfaceid": "b2ae5c23-49aa-4034-b884-b1017e034d61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.943490] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:d4:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2ae5c23-49aa-4034-b884-b1017e034d61', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.951926] env[62558]: DEBUG oslo.service.loopingcall [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.956884] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.956884] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78f44141-c7db-44c2-9e72-8aecb4ae0043 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.974637] env[62558]: DEBUG nova.compute.utils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 987.980503] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 987.980787] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 987.996030] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.996030] env[62558]: value = "task-1267188" [ 987.996030] env[62558]: _type = "Task" [ 987.996030] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.007787] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267188, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.095763] env[62558]: DEBUG nova.policy [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc5633ea3c434344b2ff745ef407a66d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a090a1a385e43dc840b63eabb74d0cc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.229459] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267186, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.240198] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267185, 'name': CloneVM_Task} progress is 94%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.270067] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe32ce7-f049-4c5c-8c2c-5515f749ff2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.280589] env[62558]: DEBUG oslo_vmware.api [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267187, 'name': PowerOnVM_Task} progress is 93%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.284261] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd51b0d-d91a-46d1-a6de-ca64a9f04938 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.327018] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094a8c46-7fed-48ac-a381-7324457ec38d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.340269] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dab6f90-89a5-469c-b694-be2d7c73b1c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.365558] env[62558]: DEBUG nova.compute.provider_tree [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.434831] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.435773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.435982] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.436940] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e722f263-e57e-489e-a191-defde0945ecb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.459626] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.460014] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.460205] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.460390] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.460538] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.460842] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.460886] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.461068] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.461273] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.461440] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.461621] env[62558]: DEBUG nova.virt.hardware [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.468644] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfiguring VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 988.469099] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a7f28e0-7394-4900-9792-74ffc3ee2faf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.482836] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 988.498873] env[62558]: DEBUG oslo_vmware.api [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 988.498873] env[62558]: value = "task-1267189" [ 988.498873] env[62558]: _type = "Task" [ 988.498873] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.514345] env[62558]: DEBUG oslo_vmware.api [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267189, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.518043] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267188, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.521530] env[62558]: DEBUG nova.compute.manager [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-changed-5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.521765] env[62558]: DEBUG nova.compute.manager [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing instance network info cache due to event network-changed-5f5f33c6-f6bc-4194-9da2-9e3bca010a13. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.522054] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.522172] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.522310] env[62558]: DEBUG nova.network.neutron [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing network info cache for port 5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.554224] env[62558]: DEBUG nova.compute.manager [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Received event network-changed-b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.554224] env[62558]: DEBUG nova.compute.manager [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Refreshing instance network info cache due to event network-changed-b2ae5c23-49aa-4034-b884-b1017e034d61. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.554224] env[62558]: DEBUG oslo_concurrency.lockutils [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] Acquiring lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.554224] env[62558]: DEBUG oslo_concurrency.lockutils [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] Acquired lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.554224] env[62558]: DEBUG nova.network.neutron [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Refreshing network info cache for port b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.557443] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Successfully created port: 5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 988.567839] env[62558]: DEBUG nova.network.neutron [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.727720] env[62558]: DEBUG oslo_vmware.api [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267186, 'name': PowerOnVM_Task, 'duration_secs': 0.724099} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.727720] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.727907] env[62558]: INFO nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Took 9.96 seconds to spawn the instance on the hypervisor. [ 988.727997] env[62558]: DEBUG nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.728872] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b673ea01-86cf-4ebe-b2cf-9ca8aefa22af {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.751610] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267185, 'name': CloneVM_Task, 'duration_secs': 1.327849} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.752327] env[62558]: INFO nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Created linked-clone VM from snapshot [ 988.753567] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2270006f-6afe-44f0-a49b-c6223f4122b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.770634] env[62558]: DEBUG nova.virt.vmwareapi.images [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Uploading image 8cff4e50-3e1d-4c5c-8437-bbe81ace6bef {{(pid=62558) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 988.793424] env[62558]: DEBUG oslo_vmware.api [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267187, 'name': PowerOnVM_Task, 'duration_secs': 0.641268} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.798070] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Destroying the VM {{(pid=62558) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 988.798070] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Resumed the VM {{(pid=62558) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 988.798070] env[62558]: DEBUG nova.compute.manager [None req-6e0a38fc-dc82-457e-9afa-3020f1d914a9 tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.798070] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ef42568c-eec6-4e35-bf1e-9f44209fb487 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.801453] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd153523-bfc6-4e07-abcb-d47b4869e6ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.818609] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 988.818609] env[62558]: value = "task-1267190" [ 988.818609] env[62558]: _type = "Task" [ 988.818609] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.830751] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267190, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.871022] env[62558]: DEBUG nova.scheduler.client.report [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.017170] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267188, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.028248] env[62558]: DEBUG oslo_vmware.api [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267189, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.070127] env[62558]: INFO nova.compute.manager [-] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Took 1.94 seconds to deallocate network for instance. [ 989.262675] env[62558]: INFO nova.compute.manager [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Took 24.35 seconds to build instance. [ 989.313716] env[62558]: DEBUG nova.network.neutron [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updated VIF entry in instance network info cache for port 5f5f33c6-f6bc-4194-9da2-9e3bca010a13. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.314342] env[62558]: DEBUG nova.network.neutron [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "address": "fa:16:3e:e0:7b:e2", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f5f33c6-f6", "ovs_interfaceid": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.331821] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267190, 'name': Destroy_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.373680] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.916s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.395875] env[62558]: DEBUG nova.network.neutron [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updated VIF entry in instance network info cache for port b2ae5c23-49aa-4034-b884-b1017e034d61. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.396595] env[62558]: DEBUG nova.network.neutron [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updating instance_info_cache with network_info: [{"id": "b2ae5c23-49aa-4034-b884-b1017e034d61", "address": "fa:16:3e:92:d4:32", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ae5c23-49", "ovs_interfaceid": "b2ae5c23-49aa-4034-b884-b1017e034d61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.399182] env[62558]: INFO nova.scheduler.client.report [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568 [ 989.494719] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 989.508028] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267188, 'name': CreateVM_Task, 'duration_secs': 1.035255} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.511113] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.511858] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.512040] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.512398] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.513015] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543ea056-98ef-4b26-82c4-96ba9f71a425 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.521257] env[62558]: DEBUG oslo_vmware.api [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267189, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.524301] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 989.524301] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da0a80-a557-528f-c002-106811243ead" [ 989.524301] env[62558]: _type = "Task" [ 989.524301] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.529227] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.529461] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.529620] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.529799] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.529969] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.530127] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.530350] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.530535] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.530712] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.530897] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.531090] env[62558]: DEBUG nova.virt.hardware [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.531844] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aeb7918-6f17-4314-bb94-239f6907ea81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.542360] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da0a80-a557-528f-c002-106811243ead, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.545475] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cacbcc-ffa8-4a59-82ec-622db1e9fb95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.576972] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.577264] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.577495] env[62558]: DEBUG nova.objects.instance [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'resources' on Instance uuid 4815ba3f-265f-466a-9850-4c325cdb88de {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.767352] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fc5e3de9-8161-4ed9-a168-d8fde526cf6a tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.865s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.788311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.788760] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.788876] env[62558]: DEBUG nova.compute.manager [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Going to confirm migration 2 {{(pid=62558) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 989.823553] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.823553] env[62558]: DEBUG nova.compute.manager [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Received event network-vif-deleted-7dba3d4c-c3d7-4a6b-b03b-df7f321245c1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.823553] env[62558]: INFO nova.compute.manager [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Neutron deleted interface 7dba3d4c-c3d7-4a6b-b03b-df7f321245c1; detaching it from the instance and deleting it from the info cache [ 989.823553] env[62558]: DEBUG nova.network.neutron [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.834376] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267190, 'name': Destroy_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.902826] env[62558]: DEBUG oslo_concurrency.lockutils [req-5b3c3842-bc7b-4e38-917d-1b2e81e4b027 req-32fff662-cce1-4605-b129-e40f3052ea90 service nova] Releasing lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.912279] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e0b1b82a-a66c-4c88-9d15-60efbc4bb521 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.279s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.025828] env[62558]: DEBUG oslo_vmware.api [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267189, 'name': ReconfigVM_Task, 'duration_secs': 1.402201} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.025828] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.025828] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfigured VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 990.038990] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52da0a80-a557-528f-c002-106811243ead, 'name': SearchDatastore_Task, 'duration_secs': 0.031223} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.040707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.041168] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.041522] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.041799] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.042143] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.042784] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72e36e02-abab-4094-bf8c-e0d038a329ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.056886] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.057263] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.058011] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b2acc70-e1db-4837-9877-942842066239 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.064744] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 990.064744] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a96697-626c-8015-2a8d-cd73f9f943c4" [ 990.064744] env[62558]: _type = "Task" [ 990.064744] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.072970] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a96697-626c-8015-2a8d-cd73f9f943c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.326640] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f73e293-3911-45af-b0ad-ce948e03d071 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.329056] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-702d46a4-eceb-4ff9-9085-3f68a2488c4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.337853] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267190, 'name': Destroy_Task, 'duration_secs': 1.043101} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.340707] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Destroyed the VM [ 990.340912] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Deleting Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 990.341265] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-16b87ef2-65e1-49fb-8c02-bc978f0ff9b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.345553] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30258463-1d06-4ace-9a5a-f22365fa4530 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.357374] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03547df1-324b-4d35-8200-9bb915140f8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.406844] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf438a9-c88b-4883-b6b1-597ece962804 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.409950] env[62558]: DEBUG nova.compute.manager [req-fd7716c7-0493-4be1-ba2e-1ac7cb90a757 req-28bc3942-d8ed-4f73-9e0e-fdb14c193fcd service nova] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Detach interface failed, port_id=7dba3d4c-c3d7-4a6b-b03b-df7f321245c1, reason: Instance 4815ba3f-265f-466a-9850-4c325cdb88de could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 990.410659] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 990.410659] env[62558]: value = "task-1267191" [ 990.410659] env[62558]: _type = "Task" [ 990.410659] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.417623] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.417797] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.417971] env[62558]: DEBUG nova.network.neutron [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.418175] env[62558]: DEBUG nova.objects.instance [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'info_cache' on Instance uuid c65dbe09-d073-48ad-a18a-8b6383c7e345 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.421281] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee11af05-e2de-40f7-957b-09cf75fd21ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.430331] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.441912] env[62558]: DEBUG nova.compute.provider_tree [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.528387] env[62558]: DEBUG oslo_concurrency.lockutils [None req-43f6b03f-8a37-49ef-bec2-d521db0098c1 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.778s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.575856] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a96697-626c-8015-2a8d-cd73f9f943c4, 'name': SearchDatastore_Task, 'duration_secs': 0.022872} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.576690] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c57e532-bd57-45b0-8e89-54a91bcb9d04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.583290] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 990.583290] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52453cfc-4819-d48f-fe99-2ee825675d72" [ 990.583290] env[62558]: _type = "Task" [ 990.583290] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.592039] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52453cfc-4819-d48f-fe99-2ee825675d72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.921432] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.944852] env[62558]: DEBUG nova.scheduler.client.report [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.985367] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Successfully updated port: 5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.094579] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52453cfc-4819-d48f-fe99-2ee825675d72, 'name': SearchDatastore_Task, 'duration_secs': 0.010724} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.094859] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.095165] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ea680004-b94a-47f9-b516-2afb95c30a51/ea680004-b94a-47f9-b516-2afb95c30a51.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.095452] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dae253e3-f24f-4469-9719-a36f91525f5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.106031] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 991.106031] env[62558]: value = "task-1267192" [ 991.106031] env[62558]: _type = "Task" [ 991.106031] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.117283] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.308219] env[62558]: DEBUG nova.compute.manager [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Received event network-vif-plugged-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.308219] env[62558]: DEBUG oslo_concurrency.lockutils [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.308219] env[62558]: DEBUG oslo_concurrency.lockutils [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.308219] env[62558]: DEBUG oslo_concurrency.lockutils [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.308219] env[62558]: DEBUG nova.compute.manager [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] No waiting events found dispatching network-vif-plugged-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.308219] env[62558]: WARNING nova.compute.manager [req-1b718932-ebe4-4ab1-8334-fde30577fe55 req-1d81dd75-40e2-41b5-92c4-5392fd3cebee service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Received unexpected event network-vif-plugged-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b for instance with vm_state building and task_state spawning. [ 991.423302] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.450565] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.873s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.483571] env[62558]: INFO nova.scheduler.client.report [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted allocations for instance 4815ba3f-265f-466a-9850-4c325cdb88de [ 991.491456] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.491456] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.493300] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.616284] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267192, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.905801] env[62558]: DEBUG nova.network.neutron [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [{"id": "46f92558-3c85-4803-aad8-7ba3157e3f67", "address": "fa:16:3e:76:e8:57", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46f92558-3c", "ovs_interfaceid": "46f92558-3c85-4803-aad8-7ba3157e3f67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.927426] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.998703] env[62558]: DEBUG oslo_concurrency.lockutils [None req-324f5382-0b23-46c4-82a8-f8f2597652c5 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "4815ba3f-265f-466a-9850-4c325cdb88de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.052200] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 992.118224] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605041} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.118520] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] ea680004-b94a-47f9-b516-2afb95c30a51/ea680004-b94a-47f9-b516-2afb95c30a51.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.118743] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.119013] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-826c31bb-ac50-4088-aa37-5f82c0800eeb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.127420] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 992.127420] env[62558]: value = "task-1267193" [ 992.127420] env[62558]: _type = "Task" [ 992.127420] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.137208] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267193, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.294597] env[62558]: DEBUG nova.network.neutron [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.412026] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-c65dbe09-d073-48ad-a18a-8b6383c7e345" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.412241] env[62558]: DEBUG nova.objects.instance [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'migration_context' on Instance uuid c65dbe09-d073-48ad-a18a-8b6383c7e345 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.425517] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.555708] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.555938] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.637657] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267193, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088934} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.637937] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.638746] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09d8399-4469-419a-9cd4-c5048128d0d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.661432] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] ea680004-b94a-47f9-b516-2afb95c30a51/ea680004-b94a-47f9-b516-2afb95c30a51.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.661903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-509ff331-739e-4b0d-9d47-3ca00b4f60ec" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.662242] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-509ff331-739e-4b0d-9d47-3ca00b4f60ec" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.662610] env[62558]: DEBUG nova.objects.instance [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.664073] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e66f8b6-01b8-45f5-a07f-18cb844a573b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.686132] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 992.686132] env[62558]: value = "task-1267194" [ 992.686132] env[62558]: _type = "Task" [ 992.686132] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.695414] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267194, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.797568] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.797929] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Instance network_info: |[{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.798419] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:6b:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f78b07ea-f425-4622-84f4-706a5d8820a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bb4afdb-999e-446a-ab6e-1bd7a8bc092b', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.806414] env[62558]: DEBUG oslo.service.loopingcall [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.806679] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.806914] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff4e153b-563c-471e-9139-e677c5b8d94a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.828583] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.828583] env[62558]: value = "task-1267195" [ 992.828583] env[62558]: _type = "Task" [ 992.828583] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.837532] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267195, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.920771] env[62558]: DEBUG nova.objects.base [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 992.924893] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14df3144-bd8b-4bf9-9e3d-cdd70391cf6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.927927] env[62558]: DEBUG oslo_vmware.api [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267191, 'name': RemoveSnapshot_Task, 'duration_secs': 2.425297} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.928537] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Deleted Snapshot of the VM instance {{(pid=62558) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 992.948213] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b2e1033-5142-4c59-8df4-d6b6e3231ae6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.955942] env[62558]: DEBUG oslo_vmware.api [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 992.955942] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a50ef-ef9c-ff92-09c0-1389a5e7cc6e" [ 992.955942] env[62558]: _type = "Task" [ 992.955942] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.967658] env[62558]: DEBUG oslo_vmware.api [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a50ef-ef9c-ff92-09c0-1389a5e7cc6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.058319] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 993.196258] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267194, 'name': ReconfigVM_Task, 'duration_secs': 0.478416} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.196584] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Reconfigured VM instance instance-00000063 to attach disk [datastore1] ea680004-b94a-47f9-b516-2afb95c30a51/ea680004-b94a-47f9-b516-2afb95c30a51.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.197246] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3a41cf8-3c90-4943-a36e-4d6a989cab99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.204569] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 993.204569] env[62558]: value = "task-1267197" [ 993.204569] env[62558]: _type = "Task" [ 993.204569] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.213523] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267197, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.321434] env[62558]: DEBUG nova.objects.instance [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.323755] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.323987] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.340671] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267195, 'name': CreateVM_Task, 'duration_secs': 0.41691} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.340834] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.341541] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.341714] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.342067] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.342361] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69281941-92a7-4ddc-ad64-cccceb58b3c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.348145] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 993.348145] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fd0a4a-4e06-a962-db03-bf111b6a43c0" [ 993.348145] env[62558]: _type = "Task" [ 993.348145] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.356688] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fd0a4a-4e06-a962-db03-bf111b6a43c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.433567] env[62558]: WARNING nova.compute.manager [None req-143665b2-7a4c-475d-87e5-8632827a47b9 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Image not found during snapshot: nova.exception.ImageNotFound: Image 8cff4e50-3e1d-4c5c-8437-bbe81ace6bef could not be found. [ 993.468301] env[62558]: DEBUG oslo_vmware.api [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521a50ef-ef9c-ff92-09c0-1389a5e7cc6e, 'name': SearchDatastore_Task, 'duration_secs': 0.010279} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.468605] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.468837] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.581403] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.715438] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267197, 'name': Rename_Task, 'duration_secs': 0.147517} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.715729] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.715976] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07d42572-2af7-4d70-b78d-9b13d210abd0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.723296] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 993.723296] env[62558]: value = "task-1267198" [ 993.723296] env[62558]: _type = "Task" [ 993.723296] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.730690] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.824331] env[62558]: DEBUG nova.objects.base [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance<29584082-37e6-4dd8-906f-fd87b4ca9bc2> lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 993.824498] env[62558]: DEBUG nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 993.827561] env[62558]: DEBUG nova.compute.utils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.859217] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fd0a4a-4e06-a962-db03-bf111b6a43c0, 'name': SearchDatastore_Task, 'duration_secs': 0.047741} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.861375] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.861622] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.861853] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.862009] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.862215] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.862484] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28e94790-443d-4383-b0f6-243051f8fd68 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.871899] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.872155] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.872983] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edf52d3f-52a3-4a9a-ac58-c9db838ec6fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.880495] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 993.880495] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc1136-00fc-66af-bec0-1375c4bed664" [ 993.880495] env[62558]: _type = "Task" [ 993.880495] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.888262] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc1136-00fc-66af-bec0-1375c4bed664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.900420] env[62558]: DEBUG nova.policy [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.165058] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8568793d-34a4-4d9e-add1-6c2e9dc0fce2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.173062] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaccee4e-1e1b-4780-bca0-7ff66e87aa0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.204087] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add8a06d-f803-4624-aaf1-c9d1a779b558 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.212500] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec94322-2af0-410a-943e-4bcea18e24e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.227069] env[62558]: DEBUG nova.compute.provider_tree [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.236264] env[62558]: DEBUG oslo_vmware.api [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267198, 'name': PowerOnVM_Task, 'duration_secs': 0.464696} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.237117] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.237335] env[62558]: INFO nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Took 8.38 seconds to spawn the instance on the hypervisor. [ 994.237518] env[62558]: DEBUG nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.238277] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6434b428-56e8-4849-8f3f-5fbe7a53298a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.331220] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.390773] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bc1136-00fc-66af-bec0-1375c4bed664, 'name': SearchDatastore_Task, 'duration_secs': 0.012558} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.391558] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2437dde9-8a17-4e6b-9231-fb0d0cf7af81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.396660] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 994.396660] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e9d-4bb1-f6bd-7cb8-7e5ba4a83d82" [ 994.396660] env[62558]: _type = "Task" [ 994.396660] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.403996] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e9d-4bb1-f6bd-7cb8-7e5ba4a83d82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.732769] env[62558]: DEBUG nova.scheduler.client.report [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.754414] env[62558]: INFO nova.compute.manager [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Took 15.40 seconds to build instance. [ 994.907894] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e9d-4bb1-f6bd-7cb8-7e5ba4a83d82, 'name': SearchDatastore_Task, 'duration_secs': 0.051964} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.908207] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.908480] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.908749] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a448186a-82b9-4009-a013-509fd2423e4e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.916254] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 994.916254] env[62558]: value = "task-1267199" [ 994.916254] env[62558]: _type = "Task" [ 994.916254] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.924878] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.012376] env[62558]: DEBUG nova.compute.manager [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Received event network-changed-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.012376] env[62558]: DEBUG nova.compute.manager [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Refreshing instance network info cache due to event network-changed-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.012501] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Acquiring lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.012629] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Acquired lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.013165] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Refreshing network info cache for port 5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 995.257025] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2ce459ff-69d9-4fcc-946a-271ff80a3a83 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.906s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.385493] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.385852] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.386171] env[62558]: INFO nova.compute.manager [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Attaching volume d479af90-d34c-43cd-a255-317c4b2c3d13 to /dev/sdb [ 995.423558] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d93519-e0fe-4661-bb15-941c4183443e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.429839] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267199, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.434842] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fccd750-2f53-48e4-833e-4eb4eb41e272 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.451154] env[62558]: DEBUG nova.virt.block_device [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating existing volume attachment record: e05607bc-0f7e-4ebf-9365-d0268aed27bf {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 995.562017] env[62558]: DEBUG nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Successfully updated port: 509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.727021] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updated VIF entry in instance network info cache for port 5bb4afdb-999e-446a-ab6e-1bd7a8bc092b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.727462] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.743726] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.275s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.747899] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.167s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.749747] env[62558]: INFO nova.compute.claims [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.927662] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53604} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.927928] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.928160] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.928415] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6878f37c-8d17-4712-a908-48971e5aed45 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.936731] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 995.936731] env[62558]: value = "task-1267203" [ 995.936731] env[62558]: _type = "Task" [ 995.936731] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.947477] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.065386] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.065642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.065799] env[62558]: DEBUG nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.230566] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Releasing lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.230953] env[62558]: DEBUG nova.compute.manager [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Received event network-changed-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.231157] env[62558]: DEBUG nova.compute.manager [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Refreshing instance network info cache due to event network-changed-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.231379] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Acquiring lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.231522] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Acquired lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.231687] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Refreshing network info cache for port c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.310368] env[62558]: INFO nova.scheduler.client.report [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocation for migration 9b9fa3c1-4591-4035-8fc3-73041d59ebdd [ 996.394099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.394099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.394099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.394099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.394099] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.399625] env[62558]: INFO nova.compute.manager [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Terminating instance [ 996.401661] env[62558]: DEBUG nova.compute.manager [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.401869] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.402795] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a01c03d-3fd1-4a14-a587-3ae23b8ab50d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.414611] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.414814] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41fb713c-014c-47d4-9ca7-903f2c004895 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.422391] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 996.422391] env[62558]: value = "task-1267204" [ 996.422391] env[62558]: _type = "Task" [ 996.422391] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.431867] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.446195] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086315} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.446310] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.447086] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5698e847-ddde-42a8-bea4-5f976d726897 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.469685] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.470015] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd7f0e75-48dd-491a-b796-2d8cc99b0e79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.494033] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 996.494033] env[62558]: value = "task-1267205" [ 996.494033] env[62558]: _type = "Task" [ 996.494033] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.502171] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267205, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.606645] env[62558]: WARNING nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 996.606906] env[62558]: WARNING nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 996.815618] env[62558]: DEBUG oslo_concurrency.lockutils [None req-27d9081b-f4a8-4f85-91a1-5b1dbb9d8d15 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.027s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.940221] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267204, 'name': PowerOffVM_Task, 'duration_secs': 0.315749} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.940535] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.940714] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.942170] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ca5d274-bd5b-43e0-8ac4-ddda36d9db60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.959728] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1040b1-6597-446f-85fa-af99894aeb50 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.968617] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f426908-9dd6-4023-a38a-b5426ee1d2de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.019816] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b2762e-4213-45fd-a593-384e0beec573 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.024310] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.028023] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.028023] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleting the datastore file [datastore2] ec2b48ff-dffa-4948-9925-a1c23aa76e64 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.028023] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ed39c02-e929-4207-8d84-d84f72129d41 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.033524] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64155f62-2079-42fa-915d-4e8817389a5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.037743] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267205, 'name': ReconfigVM_Task, 'duration_secs': 0.370588} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.039075] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.039942] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for the task: (returnval){ [ 997.039942] env[62558]: value = "task-1267207" [ 997.039942] env[62558]: _type = "Task" [ 997.039942] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.040553] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8412b3f8-5c4c-48a8-a2cd-7ead4e57656d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.055405] env[62558]: DEBUG nova.compute.provider_tree [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.062755] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267207, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.068491] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 997.068491] env[62558]: value = "task-1267208" [ 997.068491] env[62558]: _type = "Task" [ 997.068491] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.079128] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267208, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.133911] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updated VIF entry in instance network info cache for port c8e25a3a-b0f5-428b-a485-ff34df8b7fe5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.134631] env[62558]: DEBUG nova.network.neutron [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating instance_info_cache with network_info: [{"id": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "address": "fa:16:3e:25:4b:2d", "network": {"id": "890b4237-4fb4-4dbb-90f4-e1c5584f7291", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1581343725-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebd5df2c8e5f42f5b4b3da70f8793ed9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8e25a3a-b0", "ovs_interfaceid": "c8e25a3a-b0f5-428b-a485-ff34df8b7fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.145656] env[62558]: DEBUG nova.network.neutron [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "address": "fa:16:3e:e0:7b:e2", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f5f33c6-f6", "ovs_interfaceid": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "address": "fa:16:3e:22:08:0d", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap509ff331-73", "ovs_interfaceid": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.190145] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-plugged-509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.190583] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.190733] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.190910] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.191108] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] No waiting events found dispatching network-vif-plugged-509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.191266] env[62558]: WARNING nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received unexpected event network-vif-plugged-509ff331-739e-4b0d-9d47-3ca00b4f60ec for instance with vm_state active and task_state None. [ 997.191434] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-changed-509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.191589] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing instance network info cache due to event network-changed-509ff331-739e-4b0d-9d47-3ca00b4f60ec. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.191959] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.552787] env[62558]: DEBUG oslo_vmware.api [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Task: {'id': task-1267207, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200364} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.553732] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.553732] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.553732] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.553868] env[62558]: INFO nova.compute.manager [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Took 1.15 seconds to destroy the instance on the hypervisor. [ 997.554306] env[62558]: DEBUG oslo.service.loopingcall [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.554306] env[62558]: DEBUG nova.compute.manager [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.554306] env[62558]: DEBUG nova.network.neutron [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.558795] env[62558]: DEBUG nova.scheduler.client.report [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.581393] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267208, 'name': Rename_Task, 'duration_secs': 0.159726} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.581717] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.581971] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-564c6b68-c580-4c34-869b-a24fbf961ec6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.589572] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 997.589572] env[62558]: value = "task-1267209" [ 997.589572] env[62558]: _type = "Task" [ 997.589572] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.598605] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.638416] env[62558]: DEBUG oslo_concurrency.lockutils [req-b7d07a4c-59cb-4c72-8a66-18a19a12c65c req-adad448a-fb05-4a8c-b68f-8952ee25b3f4 service nova] Releasing lock "refresh_cache-4e8b3008-693d-4445-937b-8e7db13c7c91" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.649032] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.649646] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.650052] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.650158] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.650296] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Refreshing network info cache for port 509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.652124] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25fa2c4-e510-4174-92ae-eddb47db4697 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.670971] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.671181] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.671347] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.671529] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.671679] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.671830] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.672044] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.672240] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.672432] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.672604] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.672782] env[62558]: DEBUG nova.virt.hardware [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.679140] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfiguring VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 997.680103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e97a4b2-9b72-40dd-946e-a0e3e7680a32 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.699521] env[62558]: DEBUG oslo_vmware.api [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 997.699521] env[62558]: value = "task-1267210" [ 997.699521] env[62558]: _type = "Task" [ 997.699521] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.708218] env[62558]: DEBUG oslo_vmware.api [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267210, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.945744] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.946120] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.946298] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.946506] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.946718] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.949129] env[62558]: INFO nova.compute.manager [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Terminating instance [ 997.951121] env[62558]: DEBUG nova.compute.manager [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.951324] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.952320] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30182c16-359a-4121-8dcf-ddbf40f2d4d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.961683] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.962026] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb5db78a-c6a1-49b3-a4e7-0444a2478208 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.970407] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 997.970407] env[62558]: value = "task-1267212" [ 997.970407] env[62558]: _type = "Task" [ 997.970407] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.981710] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.010514] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.010936] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.011200] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.011404] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.012346] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.015670] env[62558]: INFO nova.compute.manager [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Terminating instance [ 998.019282] env[62558]: DEBUG nova.compute.manager [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.019847] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.021439] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14758bfd-ee25-42e9-a27a-d2ac301d5833 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.032965] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.033331] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa24b80b-230d-4382-af9f-882a6966da1b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.040631] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 998.040631] env[62558]: value = "task-1267213" [ 998.040631] env[62558]: _type = "Task" [ 998.040631] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.050615] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.063793] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.064470] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 998.101908] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267209, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.104913] env[62558]: DEBUG nova.compute.manager [req-7889fe21-ad4e-4012-95c1-ff4a90cfd12b req-9004816f-baa1-4543-ac72-02a2f7638da5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Received event network-vif-deleted-a703fd04-54f4-4a70-a8d2-8b588f0b5658 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.105174] env[62558]: INFO nova.compute.manager [req-7889fe21-ad4e-4012-95c1-ff4a90cfd12b req-9004816f-baa1-4543-ac72-02a2f7638da5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Neutron deleted interface a703fd04-54f4-4a70-a8d2-8b588f0b5658; detaching it from the instance and deleting it from the info cache [ 998.105441] env[62558]: DEBUG nova.network.neutron [req-7889fe21-ad4e-4012-95c1-ff4a90cfd12b req-9004816f-baa1-4543-ac72-02a2f7638da5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.211200] env[62558]: DEBUG oslo_vmware.api [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267210, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.338082] env[62558]: DEBUG nova.network.neutron [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.390926] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updated VIF entry in instance network info cache for port 509ff331-739e-4b0d-9d47-3ca00b4f60ec. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.391453] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "address": "fa:16:3e:e0:7b:e2", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f5f33c6-f6", "ovs_interfaceid": "5f5f33c6-f6bc-4194-9da2-9e3bca010a13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "address": "fa:16:3e:22:08:0d", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap509ff331-73", "ovs_interfaceid": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.480388] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267212, 'name': PowerOffVM_Task, 'duration_secs': 0.245112} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.480687] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.480861] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.481140] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b5d32d8-8314-4339-88a3-7369c1856ffc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.553867] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267213, 'name': PowerOffVM_Task, 'duration_secs': 0.240944} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.554809] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.554995] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.555308] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.555551] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.555734] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleting the datastore file [datastore1] 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.556038] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0222dbd-e095-41aa-9403-1c33ab9fefb1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.557788] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff4ec6ff-e60b-4b85-b368-e140b76aee95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.565961] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 998.565961] env[62558]: value = "task-1267215" [ 998.565961] env[62558]: _type = "Task" [ 998.565961] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.570721] env[62558]: DEBUG nova.compute.utils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.572706] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 998.572925] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.580480] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.601161] env[62558]: DEBUG oslo_vmware.api [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267209, 'name': PowerOnVM_Task, 'duration_secs': 0.610034} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.601458] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.601669] env[62558]: INFO nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Took 9.11 seconds to spawn the instance on the hypervisor. [ 998.601854] env[62558]: DEBUG nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.602668] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c367d2a5-482a-41f1-91bd-75624521bcf9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.614762] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cab959c1-2eb8-4f80-bcbd-3b0a90af6074 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.619469] env[62558]: DEBUG nova.policy [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.629097] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0aa9645-b2f9-49b2-a793-f2cd33b97e3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.642389] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.642607] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.642792] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleting the datastore file [datastore1] c65dbe09-d073-48ad-a18a-8b6383c7e345 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.643493] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8f44385-b1b5-49cf-bbb1-89ec2071606a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.649773] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 998.649773] env[62558]: value = "task-1267217" [ 998.649773] env[62558]: _type = "Task" [ 998.649773] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.659639] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267217, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.670407] env[62558]: DEBUG nova.compute.manager [req-7889fe21-ad4e-4012-95c1-ff4a90cfd12b req-9004816f-baa1-4543-ac72-02a2f7638da5 service nova] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Detach interface failed, port_id=a703fd04-54f4-4a70-a8d2-8b588f0b5658, reason: Instance ec2b48ff-dffa-4948-9925-a1c23aa76e64 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 998.674954] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "a94f43ba-da37-479c-b939-016dff83c754" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.675199] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.710801] env[62558]: DEBUG oslo_vmware.api [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267210, 'name': ReconfigVM_Task, 'duration_secs': 0.80113} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.711321] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.711535] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfigured VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 998.840630] env[62558]: INFO nova.compute.manager [-] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Took 1.29 seconds to deallocate network for instance. [ 998.880041] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Successfully created port: 6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.894754] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.895137] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Received event network-changed-b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.896035] env[62558]: DEBUG nova.compute.manager [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Refreshing instance network info cache due to event network-changed-b2ae5c23-49aa-4034-b884-b1017e034d61. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 998.896035] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Acquiring lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.896035] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Acquired lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.896035] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Refreshing network info cache for port b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.077439] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 999.080063] env[62558]: DEBUG oslo_vmware.api [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267215, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45499} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.080512] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.080691] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.080873] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.081059] env[62558]: INFO nova.compute.manager [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 999.081310] env[62558]: DEBUG oslo.service.loopingcall [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.081501] env[62558]: DEBUG nova.compute.manager [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.081591] env[62558]: DEBUG nova.network.neutron [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.123871] env[62558]: INFO nova.compute.manager [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Took 17.25 seconds to build instance. [ 999.161240] env[62558]: DEBUG oslo_vmware.api [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267217, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360204} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.161753] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.161753] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.161881] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.162018] env[62558]: INFO nova.compute.manager [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Took 1.14 seconds to destroy the instance on the hypervisor. [ 999.162284] env[62558]: DEBUG oslo.service.loopingcall [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.162741] env[62558]: DEBUG nova.compute.manager [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.162838] env[62558]: DEBUG nova.network.neutron [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.177207] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 999.216709] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b41e4e07-d659-46db-ab3b-54ecbab1fcc8 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-509ff331-739e-4b0d-9d47-3ca00b4f60ec" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.554s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.347728] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.347960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.348211] env[62558]: DEBUG nova.objects.instance [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lazy-loading 'resources' on Instance uuid ec2b48ff-dffa-4948-9925-a1c23aa76e64 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.626075] env[62558]: DEBUG oslo_concurrency.lockutils [None req-75bcf91c-0574-4ac6-a3ff-f23f1bc391ba tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.251s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.649497] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updated VIF entry in instance network info cache for port b2ae5c23-49aa-4034-b884-b1017e034d61. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.649497] env[62558]: DEBUG nova.network.neutron [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updating instance_info_cache with network_info: [{"id": "b2ae5c23-49aa-4034-b884-b1017e034d61", "address": "fa:16:3e:92:d4:32", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2ae5c23-49", "ovs_interfaceid": "b2ae5c23-49aa-4034-b884-b1017e034d61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.704688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.836210] env[62558]: DEBUG nova.compute.manager [req-8cf79d5e-63ed-4cc2-8018-5cb97c739de7 req-9bfd5139-6f49-40cc-86ed-d01e37963288 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Received event network-vif-deleted-15338668-abe5-414d-91c8-3396fdaf1b3e {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.836210] env[62558]: INFO nova.compute.manager [req-8cf79d5e-63ed-4cc2-8018-5cb97c739de7 req-9bfd5139-6f49-40cc-86ed-d01e37963288 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Neutron deleted interface 15338668-abe5-414d-91c8-3396fdaf1b3e; detaching it from the instance and deleting it from the info cache [ 999.836311] env[62558]: DEBUG nova.network.neutron [req-8cf79d5e-63ed-4cc2-8018-5cb97c739de7 req-9bfd5139-6f49-40cc-86ed-d01e37963288 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.998284] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 999.998284] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 999.999027] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f63c9a8-6b8c-4d77-a25e-d410d5df22d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.021352] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d3e944-687c-4cff-92d9-fad66583ea49 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.047210] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.050041] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c0e8a24-f843-4114-8329-06c9135f5f4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.069101] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1000.069101] env[62558]: value = "task-1267218" [ 1000.069101] env[62558]: _type = "Task" [ 1000.069101] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.078767] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267218, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.086831] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f40c00c-d46e-4692-9164-5d9b57bea421 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.090363] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1000.098150] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dcb3e7-eb72-48f1-bc2c-8f4a3cb0198b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.133336] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae380330-aeda-4793-b92c-0263d49e836f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.141996] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0137e1b3-4941-4448-95c1-b3ca34e7058d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.148082] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.148362] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.148542] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.148733] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.148898] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.149063] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.149276] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.149437] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.149608] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.149771] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.149944] env[62558]: DEBUG nova.virt.hardware [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.150725] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9400d86c-dea7-43bc-ac7f-858cfc642c71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.153831] env[62558]: DEBUG oslo_concurrency.lockutils [req-08753823-d70e-4739-8511-ab698f6b927a req-1a6a564a-e881-429f-b891-7eb6e45ce3c3 service nova] Releasing lock "refresh_cache-ea680004-b94a-47f9-b516-2afb95c30a51" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.167147] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b2e60b-8edc-42cd-8509-1ff5cc98b8f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.171589] env[62558]: DEBUG nova.compute.provider_tree [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.218325] env[62558]: DEBUG nova.network.neutron [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.284347] env[62558]: DEBUG nova.network.neutron [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.338792] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea4e09a2-d8e8-4cc2-8bb3-a27090f99bf0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.349340] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78ab35a-c6aa-4524-9266-34eebc0e0470 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.381176] env[62558]: DEBUG nova.compute.manager [req-8cf79d5e-63ed-4cc2-8018-5cb97c739de7 req-9bfd5139-6f49-40cc-86ed-d01e37963288 service nova] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Detach interface failed, port_id=15338668-abe5-414d-91c8-3396fdaf1b3e, reason: Instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1000.399137] env[62558]: DEBUG nova.compute.manager [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Received event network-vif-plugged-6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.399267] env[62558]: DEBUG oslo_concurrency.lockutils [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] Acquiring lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.399557] env[62558]: DEBUG oslo_concurrency.lockutils [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.399748] env[62558]: DEBUG oslo_concurrency.lockutils [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.399919] env[62558]: DEBUG nova.compute.manager [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] No waiting events found dispatching network-vif-plugged-6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.400092] env[62558]: WARNING nova.compute.manager [req-b88edec3-ae1a-4a56-ac17-56e5a03dbe71 req-35e56ca3-8665-42a4-b458-f7ab200e988d service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Received unexpected event network-vif-plugged-6c9df521-94b5-4fb1-a893-df919fcd1856 for instance with vm_state building and task_state spawning. [ 1000.451123] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Successfully updated port: 6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.579668] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267218, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.674976] env[62558]: DEBUG nova.scheduler.client.report [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.721010] env[62558]: INFO nova.compute.manager [-] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Took 1.64 seconds to deallocate network for instance. [ 1000.770030] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.770334] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.770558] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.770750] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.770918] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.773117] env[62558]: INFO nova.compute.manager [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Terminating instance [ 1000.775081] env[62558]: DEBUG nova.compute.manager [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.775300] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.776370] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb4bf71-6aca-40c5-aa64-ae97d02df2ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.785641] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.785900] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca6ba923-7a37-4579-ba7a-a51e0176d234 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.787636] env[62558]: INFO nova.compute.manager [-] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Took 1.62 seconds to deallocate network for instance. [ 1000.794188] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 1000.794188] env[62558]: value = "task-1267219" [ 1000.794188] env[62558]: _type = "Task" [ 1000.794188] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.803568] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.954386] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.954518] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.954666] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.080153] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267218, 'name': ReconfigVM_Task, 'duration_secs': 0.746205} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.080484] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.085580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-5f5f33c6-f6bc-4194-9da2-9e3bca010a13" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.085816] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-5f5f33c6-f6bc-4194-9da2-9e3bca010a13" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.087075] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e22641fd-2612-4041-ac30-9194da606ce9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.097722] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.097935] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.099089] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920f9e0c-8b1a-4ab3-8ca0-25b747b2313f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.119522] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2ad27d-d9c4-4fd3-8432-f76480e0a11f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.122272] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1001.122272] env[62558]: value = "task-1267220" [ 1001.122272] env[62558]: _type = "Task" [ 1001.122272] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.150677] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfiguring VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1001.151532] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8b155e3-8de3-48aa-8826-51bf0d728c68 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.168374] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.174213] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1001.174213] env[62558]: value = "task-1267221" [ 1001.174213] env[62558]: _type = "Task" [ 1001.174213] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.183256] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.185213] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.185695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.481s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.187161] env[62558]: INFO nova.compute.claims [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.211125] env[62558]: INFO nova.scheduler.client.report [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Deleted allocations for instance ec2b48ff-dffa-4948-9925-a1c23aa76e64 [ 1001.228309] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.294019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.305175] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267219, 'name': PowerOffVM_Task, 'duration_secs': 0.252463} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.305473] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.305648] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.305904] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-597744c8-3254-408b-bf54-700896303871 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.340197] env[62558]: DEBUG nova.compute.manager [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1001.434888] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.435138] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.435334] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleting the datastore file [datastore1] 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.435653] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cda30a08-1651-428b-8835-9aebdf8cb1a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.442213] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for the task: (returnval){ [ 1001.442213] env[62558]: value = "task-1267223" [ 1001.442213] env[62558]: _type = "Task" [ 1001.442213] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.450381] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.484924] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.628201] env[62558]: DEBUG nova.network.neutron [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Updating instance_info_cache with network_info: [{"id": "6c9df521-94b5-4fb1-a893-df919fcd1856", "address": "fa:16:3e:c9:47:92", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9df521-94", "ovs_interfaceid": "6c9df521-94b5-4fb1-a893-df919fcd1856", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.636048] env[62558]: DEBUG oslo_vmware.api [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267220, 'name': ReconfigVM_Task, 'duration_secs': 0.160341} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.636468] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1001.685796] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.718753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-c1f71090-0bb0-40ca-9927-ed03055f7143 tempest-ImagesTestJSON-212512879 tempest-ImagesTestJSON-212512879-project-member] Lock "ec2b48ff-dffa-4948-9925-a1c23aa76e64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.326s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.856615] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.860553] env[62558]: DEBUG nova.compute.manager [req-b002aa0a-96e2-45c9-9366-44fefe7937b2 req-e1e54e63-58a5-4565-aea2-cf6216f66679 service nova] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Received event network-vif-deleted-46f92558-3c85-4803-aad8-7ba3157e3f67 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.953092] env[62558]: DEBUG oslo_vmware.api [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Task: {'id': task-1267223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.295227} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.953345] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.953537] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.953720] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.953894] env[62558]: INFO nova.compute.manager [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1001.954165] env[62558]: DEBUG oslo.service.loopingcall [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.954389] env[62558]: DEBUG nova.compute.manager [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.954486] env[62558]: DEBUG nova.network.neutron [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.131487] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.131487] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Instance network_info: |[{"id": "6c9df521-94b5-4fb1-a893-df919fcd1856", "address": "fa:16:3e:c9:47:92", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9df521-94", "ovs_interfaceid": "6c9df521-94b5-4fb1-a893-df919fcd1856", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.132133] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:47:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c9df521-94b5-4fb1-a893-df919fcd1856', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.139440] env[62558]: DEBUG oslo.service.loopingcall [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.141855] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.142400] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f636aea8-7230-4349-acf6-782c65dfd395 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.168175] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.168175] env[62558]: value = "task-1267224" [ 1002.168175] env[62558]: _type = "Task" [ 1002.168175] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.175499] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267224, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.186735] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.400884] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0652c98a-e5c0-4a8a-aa20-0557558efb5f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.410005] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336eb45b-39c3-4595-8654-08a4b265258a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.445667] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038e5c48-c47c-4c1b-8758-1e24b818fa45 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.449633] env[62558]: DEBUG nova.compute.manager [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Received event network-changed-6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.449825] env[62558]: DEBUG nova.compute.manager [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Refreshing instance network info cache due to event network-changed-6c9df521-94b5-4fb1-a893-df919fcd1856. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.450056] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] Acquiring lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.450208] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] Acquired lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.450371] env[62558]: DEBUG nova.network.neutron [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Refreshing network info cache for port 6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.458926] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f17a94-1f6c-4e92-9170-24f34ed0df36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.475660] env[62558]: DEBUG nova.compute.provider_tree [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.676291] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267224, 'name': CreateVM_Task} progress is 25%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.685514] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.700567] env[62558]: DEBUG nova.objects.instance [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'flavor' on Instance uuid 109f67db-21cf-40f9-966e-4730bff98bcf {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.761237] env[62558]: DEBUG nova.network.neutron [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.978985] env[62558]: DEBUG nova.scheduler.client.report [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.171763] env[62558]: DEBUG nova.network.neutron [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Updated VIF entry in instance network info cache for port 6c9df521-94b5-4fb1-a893-df919fcd1856. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1003.171763] env[62558]: DEBUG nova.network.neutron [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Updating instance_info_cache with network_info: [{"id": "6c9df521-94b5-4fb1-a893-df919fcd1856", "address": "fa:16:3e:c9:47:92", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c9df521-94", "ovs_interfaceid": "6c9df521-94b5-4fb1-a893-df919fcd1856", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.185100] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267224, 'name': CreateVM_Task, 'duration_secs': 0.72885} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.186373] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.186818] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.187019] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.187352] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.187919] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e8ff51d-5a64-4b33-8c9a-df6dd3215eb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.193201] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.196872] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1003.196872] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ee0ab-5685-575e-4905-e06adfebf715" [ 1003.196872] env[62558]: _type = "Task" [ 1003.196872] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.208491] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ee0ab-5685-575e-4905-e06adfebf715, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.208947] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a26f6211-5c7f-4b24-b4ee-8242b5e85820 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.823s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.263685] env[62558]: INFO nova.compute.manager [-] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Took 1.31 seconds to deallocate network for instance. [ 1003.485048] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.485650] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.490123] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.262s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.490268] env[62558]: DEBUG nova.objects.instance [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lazy-loading 'resources' on Instance uuid 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.631042] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.631265] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1003.631546] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1003.678927] env[62558]: DEBUG oslo_concurrency.lockutils [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] Releasing lock "refresh_cache-6687a008-3193-4a43-b67c-c10e5ffcfc11" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.678927] env[62558]: DEBUG nova.compute.manager [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Received event network-vif-deleted-37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.678927] env[62558]: INFO nova.compute.manager [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Neutron deleted interface 37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a; detaching it from the instance and deleting it from the info cache [ 1003.678927] env[62558]: DEBUG nova.network.neutron [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.693752] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.711018] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520ee0ab-5685-575e-4905-e06adfebf715, 'name': SearchDatastore_Task, 'duration_secs': 0.025802} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.711590] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.711830] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.712077] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.712232] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.712451] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.712945] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3dbb1bb1-88c8-4fd6-a149-71debfb97bca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.722105] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.723125] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.723125] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5fc0329-2226-44fe-a646-4f8f54eccf53 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.728840] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1003.728840] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e4b-aaf9-334a-c98a-516535d4ed37" [ 1003.728840] env[62558]: _type = "Task" [ 1003.728840] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.739536] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e4b-aaf9-334a-c98a-516535d4ed37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.770373] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.995924] env[62558]: DEBUG nova.compute.utils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.997613] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1003.997781] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.049955] env[62558]: DEBUG nova.policy [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c5503a5bf734483af68683a8c853a71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9d4de9f22ec414d90eb8c2ed9c42d18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.187454] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b463e863-557e-4038-9e9d-ae0846c51201 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.196199] env[62558]: INFO nova.compute.manager [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Rebuilding instance [ 1004.198324] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.205876] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d931291a-c0fe-4c04-b469-434500f3589c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.234516] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a5866d-a823-4196-975e-fccbb1b10027 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.244794] env[62558]: DEBUG nova.compute.manager [req-b1614308-42fd-4716-9d0c-93f431fa5e3a req-6db9fcf4-4765-4261-880c-959c8d5b0950 service nova] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Detach interface failed, port_id=37703c4e-0543-4ae6-bb9d-d9cb4e6c8d5a, reason: Instance 49a58b46-207f-4515-b313-afcdb2d1ced3 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1004.250520] env[62558]: DEBUG nova.compute.manager [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.251375] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8498eb6-4711-4ffc-ae41-10c2bac232e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.258829] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52298e4b-aaf9-334a-c98a-516535d4ed37, 'name': SearchDatastore_Task, 'duration_secs': 0.010129} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.260606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0f8842-ef63-4a09-b398-9e216433d50a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.264958] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d02da40b-fc5e-4a36-8070-e67a44ab5617 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.273415] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1004.273415] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5238171a-a7c4-808f-66ab-5daaabe77dc2" [ 1004.273415] env[62558]: _type = "Task" [ 1004.273415] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.304024] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1da970-ecf2-4f79-8e54-1f298204f000 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.310117] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5238171a-a7c4-808f-66ab-5daaabe77dc2, 'name': SearchDatastore_Task, 'duration_secs': 0.010728} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.311400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.311967] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6687a008-3193-4a43-b67c-c10e5ffcfc11/6687a008-3193-4a43-b67c-c10e5ffcfc11.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.311967] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7936e53b-70fa-4517-b10f-d2ba0d1729b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.317448] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c880fa66-77e9-4585-b290-9aec6b4c8933 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.322600] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1004.322600] env[62558]: value = "task-1267225" [ 1004.322600] env[62558]: _type = "Task" [ 1004.322600] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.334067] env[62558]: DEBUG nova.compute.provider_tree [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.340958] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267225, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.392548] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Successfully created port: 236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.501089] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.631351] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1004.631696] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1004.693226] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.772034] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.772381] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-470947cb-ab4b-4ffd-828b-741d6f5166bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.780887] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1004.780887] env[62558]: value = "task-1267226" [ 1004.780887] env[62558]: _type = "Task" [ 1004.780887] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.789749] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.833776] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267225, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477309} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.834055] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 6687a008-3193-4a43-b67c-c10e5ffcfc11/6687a008-3193-4a43-b67c-c10e5ffcfc11.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.834274] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.834552] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59fc42b1-5281-4a0e-937b-a662a46f011e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.836981] env[62558]: DEBUG nova.scheduler.client.report [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.845494] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1004.845494] env[62558]: value = "task-1267227" [ 1004.845494] env[62558]: _type = "Task" [ 1004.845494] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.854904] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267227, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.150751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.150911] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.151070] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1005.191263] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.292029] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267226, 'name': PowerOffVM_Task, 'duration_secs': 0.391082} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.292029] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.341480] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.343883] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.050s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.344618] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.346170] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.490s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.349201] env[62558]: INFO nova.compute.manager [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Detaching volume d479af90-d34c-43cd-a255-317c4b2c3d13 [ 1005.362581] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267227, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063122} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.363639] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.364533] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b6495b-81e2-4a3c-8cd4-b97b1aa4ac79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.367901] env[62558]: INFO nova.scheduler.client.report [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocations for instance c65dbe09-d073-48ad-a18a-8b6383c7e345 [ 1005.369722] env[62558]: INFO nova.scheduler.client.report [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted allocations for instance 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1 [ 1005.394763] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 6687a008-3193-4a43-b67c-c10e5ffcfc11/6687a008-3193-4a43-b67c-c10e5ffcfc11.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.396976] env[62558]: INFO nova.virt.block_device [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Attempting to driver detach volume d479af90-d34c-43cd-a255-317c4b2c3d13 from mountpoint /dev/sdb [ 1005.396976] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1005.396976] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1005.397182] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41aade27-5169-4585-a68b-4c28950ad547 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.413047] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0eddf6-01ea-49db-a866-65198e44dd5d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.437167] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890d1699-e6fa-4a4e-8587-85eb5134d5b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.440131] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1005.440131] env[62558]: value = "task-1267228" [ 1005.440131] env[62558]: _type = "Task" [ 1005.440131] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.446460] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4d06c0-f205-4c40-930d-579b6cbab547 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.452176] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267228, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.473049] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ef5e5e-7a4d-4aee-b4f9-d672f613c98a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.488298] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] The volume has not been displaced from its original location: [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1005.493440] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.494015] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f23c837-5c55-435c-82de-0a15cc8e19d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.510931] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.514363] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1005.514363] env[62558]: value = "task-1267229" [ 1005.514363] env[62558]: _type = "Task" [ 1005.514363] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.522932] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267229, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.537372] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.537724] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.537942] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.538227] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.538428] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.538668] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.538945] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.539200] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.539446] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.539695] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.540010] env[62558]: DEBUG nova.virt.hardware [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.541015] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8309c85-cc9f-4916-86ee-56235ed1fa9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.549758] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb28dc3-757d-4637-bcc8-d4051f32d976 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.671390] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.692442] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.853955] env[62558]: INFO nova.compute.claims [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.870350] env[62558]: DEBUG nova.compute.manager [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Received event network-vif-plugged-236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.870923] env[62558]: DEBUG oslo_concurrency.lockutils [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] Acquiring lock "a94f43ba-da37-479c-b939-016dff83c754-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.871244] env[62558]: DEBUG oslo_concurrency.lockutils [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] Lock "a94f43ba-da37-479c-b939-016dff83c754-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.871382] env[62558]: DEBUG oslo_concurrency.lockutils [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] Lock "a94f43ba-da37-479c-b939-016dff83c754-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.871576] env[62558]: DEBUG nova.compute.manager [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] No waiting events found dispatching network-vif-plugged-236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1005.871771] env[62558]: WARNING nova.compute.manager [req-43201d7b-8686-40ba-90e1-68deaf080113 req-8296374c-8680-4761-9e1e-9d4cae267e90 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Received unexpected event network-vif-plugged-236df18f-ee35-44b7-80eb-4bb9408b4ae2 for instance with vm_state building and task_state spawning. [ 1005.876029] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f9208bd-6487-4201-8d8c-3f1634642c9e tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "c65dbe09-d073-48ad-a18a-8b6383c7e345" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.865s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.881345] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ecd9528-c3e3-46fa-b38d-c561433585d0 tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.933s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.918406] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Successfully updated port: 236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1005.951754] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267228, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.028733] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267229, 'name': ReconfigVM_Task, 'duration_secs': 0.498928} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.029205] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.037103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82ed9958-cc62-4b64-af74-efee9fc42e18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.061603] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1006.061603] env[62558]: value = "task-1267230" [ 1006.061603] env[62558]: _type = "Task" [ 1006.061603] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.079268] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267230, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.194166] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.246556] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.359912] env[62558]: INFO nova.compute.resource_tracker [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating resource usage from migration 24d51b19-2879-4b7f-a7de-4e4f1b53d301 [ 1006.421621] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.421775] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.421922] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.453696] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267228, 'name': ReconfigVM_Task, 'duration_secs': 0.656405} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.456356] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 6687a008-3193-4a43-b67c-c10e5ffcfc11/6687a008-3193-4a43-b67c-c10e5ffcfc11.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.457180] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e16c0531-ff1c-4fc8-94c3-aa1a3b13cde5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.463693] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1006.463693] env[62558]: value = "task-1267231" [ 1006.463693] env[62558]: _type = "Task" [ 1006.463693] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.475297] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267231, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.528418] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ad7a3d-7ca3-40c0-b579-3a9a11231d81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.536910] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fed5cf-208c-486e-931a-a6164230cd6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.569879] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6c8696-c0b2-494c-95f2-3fc5a24b0435 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.577722] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267230, 'name': ReconfigVM_Task, 'duration_secs': 0.391973} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.579749] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1006.582932] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad7089b-e9b1-431d-9fbd-9efe5b147a86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.600246] env[62558]: DEBUG nova.compute.provider_tree [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.693335] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.749467] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.749712] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1006.749964] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.750150] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.750308] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.750458] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.839697] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "2ac801d7-af70-46e0-88b3-02caee13497d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.839886] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.840125] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.840318] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.840488] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.843343] env[62558]: INFO nova.compute.manager [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Terminating instance [ 1006.845320] env[62558]: DEBUG nova.compute.manager [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1006.845546] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.846364] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f546bc3-f5a5-4516-ae6c-a350947930d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.854280] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.854568] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81543b03-aa20-4158-9e81-21a0ff25f610 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.862824] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 1006.862824] env[62558]: value = "task-1267232" [ 1006.862824] env[62558]: _type = "Task" [ 1006.862824] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.871570] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267232, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.956034] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1006.974756] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267231, 'name': Rename_Task, 'duration_secs': 0.151898} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.977522] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.977866] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a95dc9fd-1a77-45c9-8271-56b995a109e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.986019] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1006.986019] env[62558]: value = "task-1267233" [ 1006.986019] env[62558]: _type = "Task" [ 1006.986019] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.994948] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.104867] env[62558]: DEBUG nova.scheduler.client.report [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.193766] env[62558]: DEBUG oslo_vmware.api [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267221, 'name': ReconfigVM_Task, 'duration_secs': 5.76441} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.194121] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.194368] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Reconfigured VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1007.253624] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.292011] env[62558]: DEBUG nova.network.neutron [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Updating instance_info_cache with network_info: [{"id": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "address": "fa:16:3e:14:35:af", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap236df18f-ee", "ovs_interfaceid": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.373682] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267232, 'name': PowerOffVM_Task, 'duration_secs': 0.196268} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.373891] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.374081] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.374345] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e55a46da-5085-4bac-947e-eb0061f8170b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.447471] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.447471] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.447471] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleting the datastore file [datastore1] 2ac801d7-af70-46e0-88b3-02caee13497d {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.447471] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6f2e329-6ee6-45d8-a2c8-75876eeb7a64 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.454980] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for the task: (returnval){ [ 1007.454980] env[62558]: value = "task-1267235" [ 1007.454980] env[62558]: _type = "Task" [ 1007.454980] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.466566] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267235, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.482258] env[62558]: DEBUG nova.compute.manager [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-deleted-5f5f33c6-f6bc-4194-9da2-9e3bca010a13 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.482592] env[62558]: INFO nova.compute.manager [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Neutron deleted interface 5f5f33c6-f6bc-4194-9da2-9e3bca010a13; detaching it from the instance and deleting it from the info cache [ 1007.482964] env[62558]: DEBUG nova.network.neutron [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "address": "fa:16:3e:22:08:0d", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap509ff331-73", "ovs_interfaceid": "509ff331-739e-4b0d-9d47-3ca00b4f60ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.496759] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267233, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.609753] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.263s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.609753] env[62558]: INFO nova.compute.manager [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Migrating [ 1007.617153] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.847s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.617153] env[62558]: DEBUG nova.objects.instance [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lazy-loading 'resources' on Instance uuid 49a58b46-207f-4515-b313-afcdb2d1ced3 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.636950] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.637380] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ac4bb2e-1a3e-46a1-82ab-5454c6b2c6b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.646189] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1007.646189] env[62558]: value = "task-1267236" [ 1007.646189] env[62558]: _type = "Task" [ 1007.646189] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.660032] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] VM already powered off {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1007.660356] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1007.660612] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1007.661370] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a51daf-7362-4ce6-9b77-19f75a79e933 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.680668] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc74754-b0c6-4b2c-b9cc-19b0a60a4237 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.687169] env[62558]: WARNING nova.virt.vmwareapi.driver [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1007.687466] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.688378] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88482ddc-9908-44a5-b1cd-e5d9bfe2e848 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.694722] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.694940] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3d18aed-1fa9-4a4a-8767-995d70444846 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.755376] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.755653] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.755845] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore2] 109f67db-21cf-40f9-966e-4730bff98bcf {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.756146] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77ba836e-e290-47a5-a975-b3770220c3a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.764048] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1007.764048] env[62558]: value = "task-1267238" [ 1007.764048] env[62558]: _type = "Task" [ 1007.764048] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.772280] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.795388] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.795793] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Instance network_info: |[{"id": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "address": "fa:16:3e:14:35:af", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap236df18f-ee", "ovs_interfaceid": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.796249] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:35:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f256cfee-512d-4192-9aca-6750fdb1cd4c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '236df18f-ee35-44b7-80eb-4bb9408b4ae2', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.803976] env[62558]: DEBUG oslo.service.loopingcall [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.804221] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.804491] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3d29613-ee75-4c51-b6ae-0a838e5fdb8d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.824999] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.824999] env[62558]: value = "task-1267239" [ 1007.824999] env[62558]: _type = "Task" [ 1007.824999] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.832720] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267239, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.897523] env[62558]: DEBUG nova.compute.manager [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Received event network-changed-236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.897797] env[62558]: DEBUG nova.compute.manager [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Refreshing instance network info cache due to event network-changed-236df18f-ee35-44b7-80eb-4bb9408b4ae2. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1007.898170] env[62558]: DEBUG oslo_concurrency.lockutils [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] Acquiring lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.898385] env[62558]: DEBUG oslo_concurrency.lockutils [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] Acquired lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.898561] env[62558]: DEBUG nova.network.neutron [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Refreshing network info cache for port 236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.967153] env[62558]: DEBUG oslo_vmware.api [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Task: {'id': task-1267235, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196345} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.967153] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1007.967153] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1007.967153] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.967724] env[62558]: INFO nova.compute.manager [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1007.967724] env[62558]: DEBUG oslo.service.loopingcall [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.967724] env[62558]: DEBUG nova.compute.manager [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1007.967814] env[62558]: DEBUG nova.network.neutron [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1007.986363] env[62558]: DEBUG oslo_concurrency.lockutils [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.986704] env[62558]: DEBUG oslo_concurrency.lockutils [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] Acquired lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.987760] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed79b611-6699-42ee-bff2-41df6e9a243b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.000678] env[62558]: DEBUG oslo_vmware.api [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267233, 'name': PowerOnVM_Task, 'duration_secs': 0.63556} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.016573] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.016739] env[62558]: INFO nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1008.017026] env[62558]: DEBUG nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.017755] env[62558]: DEBUG oslo_concurrency.lockutils [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] Releasing lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.018230] env[62558]: WARNING nova.compute.manager [req-8fcc30de-9f4d-49af-84c4-5162084c1b18 req-33bc9717-610d-469f-b415-79bfa96611ad service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Detach interface failed, port_id=5f5f33c6-f6bc-4194-9da2-9e3bca010a13, reason: No device with interface-id 5f5f33c6-f6bc-4194-9da2-9e3bca010a13 exists on VM: nova.exception.NotFound: No device with interface-id 5f5f33c6-f6bc-4194-9da2-9e3bca010a13 exists on VM [ 1008.021443] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7235017-c323-479f-b61f-307acb2022e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.132118] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.132583] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.132583] env[62558]: DEBUG nova.network.neutron [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.279719] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27668} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.280024] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.280992] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.280992] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.316590] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37dcd2ce-49fe-486c-a810-dede65ef956e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.332007] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab45219f-1666-484d-a92b-8605810fb63c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.339500] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267239, 'name': CreateVM_Task, 'duration_secs': 0.471542} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.340035] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.340758] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.340932] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.341272] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.341530] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df986b3f-4dea-479c-a27c-6f1faa7afc00 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.371525] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d45f679-8ce6-4da7-820e-8c23a3f9fc51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.375904] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1008.375904] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528b6d15-019b-09ea-fb9f-54eecef24b48" [ 1008.375904] env[62558]: _type = "Task" [ 1008.375904] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.383557] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e85f44b-feca-4e35-85b9-4a9b6ed80574 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.392577] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528b6d15-019b-09ea-fb9f-54eecef24b48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.405378] env[62558]: DEBUG nova.compute.provider_tree [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.468092] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.468332] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.468547] env[62558]: DEBUG nova.network.neutron [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.544432] env[62558]: INFO nova.compute.manager [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Took 14.98 seconds to build instance. [ 1008.717733] env[62558]: DEBUG nova.network.neutron [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Updated VIF entry in instance network info cache for port 236df18f-ee35-44b7-80eb-4bb9408b4ae2. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.718112] env[62558]: DEBUG nova.network.neutron [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Updating instance_info_cache with network_info: [{"id": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "address": "fa:16:3e:14:35:af", "network": {"id": "d0896edf-1f2c-41cd-bbd6-de0e9692afb2", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-757001244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9d4de9f22ec414d90eb8c2ed9c42d18", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f256cfee-512d-4192-9aca-6750fdb1cd4c", "external-id": "nsx-vlan-transportzone-821", "segmentation_id": 821, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap236df18f-ee", "ovs_interfaceid": "236df18f-ee35-44b7-80eb-4bb9408b4ae2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.785654] env[62558]: INFO nova.virt.block_device [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Booting with volume d479af90-d34c-43cd-a255-317c4b2c3d13 at /dev/sdb [ 1008.810987] env[62558]: DEBUG nova.network.neutron [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.819935] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d879ea11-a2f8-43de-a72a-e400eeb90236 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.832019] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e7d9cc-bc1a-455e-ae33-501b64312444 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.843649] env[62558]: DEBUG nova.network.neutron [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.863748] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f3e11c06-fdc5-4d55-bae4-f9ded0180cb3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.872498] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df1bd4e-d1c9-4bcd-9ef6-8ecee3159285 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.893245] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528b6d15-019b-09ea-fb9f-54eecef24b48, 'name': SearchDatastore_Task, 'duration_secs': 0.014538} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.893615] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.893936] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.894303] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.894565] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.894867] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.895249] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f86c42df-a4a9-4eea-bdeb-5d69910971b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.910383] env[62558]: DEBUG nova.scheduler.client.report [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.915022] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393d7539-180b-4c40-bb14-befc51f69eb2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.917313] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.917492] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.918369] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d0b77a9-3fb2-447a-b498-c4c699145c4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.926643] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e89b68-2129-4bbc-90c0-4b722305e650 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.929186] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1008.929186] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52654f11-5d0c-6f92-80e6-6083aa4a11c9" [ 1008.929186] env[62558]: _type = "Task" [ 1008.929186] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.938366] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52654f11-5d0c-6f92-80e6-6083aa4a11c9, 'name': SearchDatastore_Task, 'duration_secs': 0.011529} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.939123] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0bfa73b-ae26-48d0-aab5-11a101be856e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.946900] env[62558]: DEBUG nova.virt.block_device [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating existing volume attachment record: 37cca107-aa9f-46bf-8d51-3a067b109b01 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1008.952753] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1008.952753] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524fb911-7334-c8db-dbdd-e12f016d2613" [ 1008.952753] env[62558]: _type = "Task" [ 1008.952753] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.960611] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524fb911-7334-c8db-dbdd-e12f016d2613, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.048046] env[62558]: DEBUG oslo_concurrency.lockutils [None req-69d05bfd-56f2-427a-b76f-3aec0e0a3abc tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.492s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.076701] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.076928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.077169] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.077363] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.077532] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.079763] env[62558]: INFO nova.compute.manager [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Terminating instance [ 1009.083615] env[62558]: DEBUG nova.compute.manager [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.083817] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.084961] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888dd312-e7dc-4947-b8ff-691cd17019c6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.093120] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.093367] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72b32d9a-097f-4a6c-9f52-6bfd46d40de7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.106255] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1009.106255] env[62558]: value = "task-1267240" [ 1009.106255] env[62558]: _type = "Task" [ 1009.106255] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.117518] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.210461] env[62558]: INFO nova.network.neutron [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Port 509ff331-739e-4b0d-9d47-3ca00b4f60ec from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1009.210889] env[62558]: DEBUG nova.network.neutron [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [{"id": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "address": "fa:16:3e:a0:be:7c", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b4d7b4-ef", "ovs_interfaceid": "01b4d7b4-efd1-4cc4-aa28-2663beeb66be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.220454] env[62558]: DEBUG oslo_concurrency.lockutils [req-c56bd57e-05c8-499c-81ed-f1de622a5f13 req-4a276a9c-0eb3-44f0-bada-3433d342ebb0 service nova] Releasing lock "refresh_cache-a94f43ba-da37-479c-b939-016dff83c754" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.314286] env[62558]: INFO nova.compute.manager [-] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Took 1.35 seconds to deallocate network for instance. [ 1009.347093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.419808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.422172] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.169s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.422386] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.422543] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1009.423431] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7073cecb-8bf3-495e-9524-f3e8dd358570 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.432820] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cafd391-4b52-402b-8057-53a013b6c106 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.450050] env[62558]: INFO nova.scheduler.client.report [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Deleted allocations for instance 49a58b46-207f-4515-b313-afcdb2d1ced3 [ 1009.451652] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dc876a-d915-4c38-b990-5031f18dee55 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.467384] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d44ecd2-4c69-4bb7-bc74-58ecff2255e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.470395] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]524fb911-7334-c8db-dbdd-e12f016d2613, 'name': SearchDatastore_Task, 'duration_secs': 0.012198} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.470655] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.470917] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a94f43ba-da37-479c-b939-016dff83c754/a94f43ba-da37-479c-b939-016dff83c754.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1009.471533] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f117c4a3-628b-4460-ad79-a35cc6782975 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.498619] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179455MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1009.498785] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.499000] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.503504] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1009.503504] env[62558]: value = "task-1267241" [ 1009.503504] env[62558]: _type = "Task" [ 1009.503504] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.509718] env[62558]: DEBUG nova.compute.manager [req-d73ebffd-f335-4a3f-9282-8e939ca57f9f req-aac4d49f-d0a8-4f17-89b7-f918cabeb171 service nova] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Received event network-vif-deleted-d9a83e75-1cf0-4747-90ee-2fec31fff5ad {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.517152] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.617016] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267240, 'name': PowerOffVM_Task, 'duration_secs': 0.204914} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.617433] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.617662] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.618262] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec5360b3-f249-4e67-9076-49c72926027b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.623160] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.623453] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.623685] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.623898] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.624087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.626342] env[62558]: INFO nova.compute.manager [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Terminating instance [ 1009.629836] env[62558]: DEBUG nova.compute.manager [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.630069] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.631137] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eed3898-61ca-4217-8d1b-6f1cc6c68baa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.638897] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.639103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3a137cf-41ce-4e43-a92b-66882f2eb4fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.646495] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1009.646495] env[62558]: value = "task-1267243" [ 1009.646495] env[62558]: _type = "Task" [ 1009.646495] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.655174] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.713743] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-29584082-37e6-4dd8-906f-fd87b4ca9bc2" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.739372] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.739570] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.739797] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleting the datastore file [datastore1] 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.740493] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e1f5820-417d-4a5c-a8ad-5814f41c4588 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.749616] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1009.749616] env[62558]: value = "task-1267244" [ 1009.749616] env[62558]: _type = "Task" [ 1009.749616] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.762682] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267244, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.821179] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.928656] env[62558]: DEBUG nova.compute.manager [req-0a88c2ba-cab9-45f4-887b-bf210b69a247 req-b98b9fdd-336d-4eab-b370-167910d0f20e service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-deleted-509ff331-739e-4b0d-9d47-3ca00b4f60ec {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.964163] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a23f6f60-1d6f-41ec-91d8-cc86a2d73e0b tempest-ServersNegativeTestJSON-1053867914 tempest-ServersNegativeTestJSON-1053867914-project-member] Lock "49a58b46-207f-4515-b313-afcdb2d1ced3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.194s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.023397] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267241, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513563} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.023812] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] a94f43ba-da37-479c-b939-016dff83c754/a94f43ba-da37-479c-b939-016dff83c754.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1010.023913] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.024163] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06e09bcc-13b4-4c21-9c8a-b38498aa78d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.031484] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1010.031484] env[62558]: value = "task-1267245" [ 1010.031484] env[62558]: _type = "Task" [ 1010.031484] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.042339] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267245, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.157856] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267243, 'name': PowerOffVM_Task, 'duration_secs': 0.342849} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.158138] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.158318] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.158584] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96bed503-224b-46c4-8703-66cc01c79e1a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.218734] env[62558]: DEBUG oslo_concurrency.lockutils [None req-761fd28f-49d8-4eca-8aa1-7b5f0c36efec tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-29584082-37e6-4dd8-906f-fd87b4ca9bc2-5f5f33c6-f6bc-4194-9da2-9e3bca010a13" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.133s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.259610] env[62558]: DEBUG oslo_vmware.api [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267244, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.438981} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.259996] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.260277] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.260539] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.260804] env[62558]: INFO nova.compute.manager [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1010.261122] env[62558]: DEBUG oslo.service.loopingcall [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.261357] env[62558]: DEBUG nova.compute.manager [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.261441] env[62558]: DEBUG nova.network.neutron [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.335324] env[62558]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 509ff331-739e-4b0d-9d47-3ca00b4f60ec could not be found.", "detail": ""}} {{(pid=62558) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1010.335635] env[62558]: DEBUG nova.network.neutron [-] Unable to show port 509ff331-739e-4b0d-9d47-3ca00b4f60ec as it no longer exists. {{(pid=62558) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1010.435650] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.436950] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.436950] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore1] 6687a008-3193-4a43-b67c-c10e5ffcfc11 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.436950] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fadcb97-cd97-4526-9d01-2a60008cbad9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.445620] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1010.445620] env[62558]: value = "task-1267247" [ 1010.445620] env[62558]: _type = "Task" [ 1010.445620] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.458304] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.518361] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Applying migration context for instance 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 as it has an incoming, in-progress migration 24d51b19-2879-4b7f-a7de-4e4f1b53d301. Migration status is migrating {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1010.519718] env[62558]: INFO nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating resource usage from migration 24d51b19-2879-4b7f-a7de-4e4f1b53d301 [ 1010.541327] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267245, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068947} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.542232] env[62558]: WARNING nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 2ac801d7-af70-46e0-88b3-02caee13497d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1010.542376] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.542502] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a3f42034-3d2b-472a-89c0-5445cb6fb567 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.542620] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 29584082-37e6-4dd8-906f-fd87b4ca9bc2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.542742] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 109f67db-21cf-40f9-966e-4730bff98bcf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.542850] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e8b3008-693d-4445-937b-8e7db13c7c91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.542961] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ea680004-b94a-47f9-b516-2afb95c30a51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.543087] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 6687a008-3193-4a43-b67c-c10e5ffcfc11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.543200] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance a94f43ba-da37-479c-b939-016dff83c754 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.543308] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Migration 24d51b19-2879-4b7f-a7de-4e4f1b53d301 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1010.543417] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1010.543606] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1010.543745] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1010.545923] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.549099] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784a9aa0-149d-49be-97f0-fd54f69b01a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.574347] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] a94f43ba-da37-479c-b939-016dff83c754/a94f43ba-da37-479c-b939-016dff83c754.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.576539] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2381e43e-73ad-4aa8-9e22-a4fb454950b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.598814] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1010.598814] env[62558]: value = "task-1267248" [ 1010.598814] env[62558]: _type = "Task" [ 1010.598814] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.610908] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.736334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebfa7fd-9393-4c91-b0b1-30f6bc5b1ec5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.745289] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a585d2c-6fa0-483d-9819-b7101f834139 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.780100] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b2944c-e438-42aa-877d-804ab54c79b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.789756] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2651890-c85f-4c2e-8b1b-16e9b2c11ea4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.802363] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.861268] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b14775a-4b16-466e-8213-dad5aa64e799 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.882533] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1010.957915] env[62558]: DEBUG oslo_vmware.api [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157081} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.957915] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.957915] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.957915] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.957915] env[62558]: INFO nova.compute.manager [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Took 1.33 seconds to destroy the instance on the hypervisor. [ 1010.957915] env[62558]: DEBUG oslo.service.loopingcall [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.957915] env[62558]: DEBUG nova.compute.manager [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1010.957915] env[62558]: DEBUG nova.network.neutron [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.084978] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.085681] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.086057] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.086473] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.086769] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.087316] env[62558]: DEBUG nova.virt.hardware [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.088502] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5fbe87-e85d-4d6f-bcec-6b3aa9a59e9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.100417] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a0c85e-c035-4629-bb85-f9e82d3ed4e8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.124923] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:de:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94d89474-fbcc-4208-aa16-91863dd3e2e5', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.134074] env[62558]: DEBUG oslo.service.loopingcall [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.138577] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.139259] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267248, 'name': ReconfigVM_Task, 'duration_secs': 0.284962} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.139750] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d95a13e4-22a3-43d3-aa83-3d30c72f5326 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.158292] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Reconfigured VM instance instance-00000066 to attach disk [datastore2] a94f43ba-da37-479c-b939-016dff83c754/a94f43ba-da37-479c-b939-016dff83c754.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1011.159017] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-905e815d-1ad3-4cad-9d86-b442ca6df87d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.166616] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.166616] env[62558]: value = "task-1267250" [ 1011.166616] env[62558]: _type = "Task" [ 1011.166616] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.168246] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1011.168246] env[62558]: value = "task-1267249" [ 1011.168246] env[62558]: _type = "Task" [ 1011.168246] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.181471] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267249, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.185384] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267250, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.330478] env[62558]: ERROR nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [req-36d053c3-fc4a-4b5b-9572-be9cbb264765] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-36d053c3-fc4a-4b5b-9572-be9cbb264765"}]} [ 1011.353033] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1011.369800] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1011.370008] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.385282] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1011.390229] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.391029] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34b2a2eb-8389-4a49-9294-9a246e735f96 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.402601] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1011.402601] env[62558]: value = "task-1267251" [ 1011.402601] env[62558]: _type = "Task" [ 1011.402601] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.414728] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.415747] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1011.463751] env[62558]: DEBUG nova.network.neutron [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.571492] env[62558]: DEBUG nova.compute.manager [req-e78688af-5d04-42ff-9231-6209eb5b5282 req-82b4dc47-1589-4b92-b35d-a922ae07cb5c service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Received event network-vif-deleted-6c9df521-94b5-4fb1-a893-df919fcd1856 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.572433] env[62558]: INFO nova.compute.manager [req-e78688af-5d04-42ff-9231-6209eb5b5282 req-82b4dc47-1589-4b92-b35d-a922ae07cb5c service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Neutron deleted interface 6c9df521-94b5-4fb1-a893-df919fcd1856; detaching it from the instance and deleting it from the info cache [ 1011.572433] env[62558]: DEBUG nova.network.neutron [req-e78688af-5d04-42ff-9231-6209eb5b5282 req-82b4dc47-1589-4b92-b35d-a922ae07cb5c service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.637085] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3f5869-ed3a-4bb8-bdd0-2a8b415e8ca4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.646265] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea21804-044a-47c3-a725-a9ea9b46339f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.691244] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395abd60-369a-4650-9bc7-c105c278da8e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.700475] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267249, 'name': Rename_Task, 'duration_secs': 0.164681} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.705927] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.706589] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267250, 'name': CreateVM_Task, 'duration_secs': 0.430142} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.706814] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2415b83-ca90-4ac0-ae91-5f55754de259 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.709530] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa869b9-410b-4f1b-8a13-44776f0ee310 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.713887] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.714749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.715071] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.715351] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.716331] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94d23142-c9c3-4ef7-8e45-fa1847e4cf57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.728066] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.731705] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1011.731705] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522165e6-8b97-540a-2aa1-dd71ce9ce069" [ 1011.731705] env[62558]: _type = "Task" [ 1011.731705] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.732095] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1011.732095] env[62558]: value = "task-1267252" [ 1011.732095] env[62558]: _type = "Task" [ 1011.732095] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.745150] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.749315] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]522165e6-8b97-540a-2aa1-dd71ce9ce069, 'name': SearchDatastore_Task, 'duration_secs': 0.012818} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.749705] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.749882] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.750188] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.750291] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.750466] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.750717] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-adc27d51-ed9c-4111-94af-2477ade4d900 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.760589] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.760798] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.761901] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f706a0f8-b817-4c69-91cb-b44b2b65a2e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.769210] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1011.769210] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bd3f44-dc93-8bce-cf78-a13e0c024210" [ 1011.769210] env[62558]: _type = "Task" [ 1011.769210] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.780305] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bd3f44-dc93-8bce-cf78-a13e0c024210, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.915719] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267251, 'name': PowerOffVM_Task, 'duration_secs': 0.367243} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.916083] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.916357] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1011.920234] env[62558]: DEBUG nova.network.neutron [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.966880] env[62558]: INFO nova.compute.manager [-] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Took 1.71 seconds to deallocate network for instance. [ 1012.059141] env[62558]: DEBUG nova.compute.manager [req-247b58f8-9149-401c-bed5-ef81a7a79aba req-de857969-1af9-4e3e-8368-e523de95a4aa service nova] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Received event network-vif-deleted-01b4d7b4-efd1-4cc4-aa28-2663beeb66be {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.074099] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-319ed453-47e2-4685-8ff4-3740aded0f25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.086144] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e859c335-0a49-4638-a700-0254f8c1bb3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.119402] env[62558]: DEBUG nova.compute.manager [req-e78688af-5d04-42ff-9231-6209eb5b5282 req-82b4dc47-1589-4b92-b35d-a922ae07cb5c service nova] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Detach interface failed, port_id=6c9df521-94b5-4fb1-a893-df919fcd1856, reason: Instance 6687a008-3193-4a43-b67c-c10e5ffcfc11 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1012.246029] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.269161] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1012.269364] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 124 to 125 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1012.269524] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.284374] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bd3f44-dc93-8bce-cf78-a13e0c024210, 'name': SearchDatastore_Task, 'duration_secs': 0.013626} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.285856] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4243d717-f72a-473a-a286-eab3cd4b508d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.291935] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1012.291935] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5288664d-11d6-3ccd-e213-fc1ec1322219" [ 1012.291935] env[62558]: _type = "Task" [ 1012.291935] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.299788] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5288664d-11d6-3ccd-e213-fc1ec1322219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.424015] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.424392] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.424515] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.424892] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.425180] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.425447] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.425911] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.426347] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.426347] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.426522] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.426723] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.431824] env[62558]: INFO nova.compute.manager [-] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Took 1.47 seconds to deallocate network for instance. [ 1012.432103] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d2bbb09-8a65-4ac3-977b-8fd681193f0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.452430] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1012.452430] env[62558]: value = "task-1267253" [ 1012.452430] env[62558]: _type = "Task" [ 1012.452430] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.464995] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.474820] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.750187] env[62558]: DEBUG oslo_vmware.api [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267252, 'name': PowerOnVM_Task, 'duration_secs': 0.76732} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.750661] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.751119] env[62558]: INFO nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Took 7.24 seconds to spawn the instance on the hypervisor. [ 1012.751478] env[62558]: DEBUG nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.752706] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e3cf02-56dd-49ec-be0a-9c45c8497932 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.777412] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1012.777707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.279s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.778112] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.957s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.778392] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.788209] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.312s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.788584] env[62558]: DEBUG nova.objects.instance [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'resources' on Instance uuid 29584082-37e6-4dd8-906f-fd87b4ca9bc2 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.805988] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5288664d-11d6-3ccd-e213-fc1ec1322219, 'name': SearchDatastore_Task, 'duration_secs': 0.01306} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.808320] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.808320] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.808320] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-213d49aa-1a7a-476f-895f-d0404eeadc04 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.815868] env[62558]: INFO nova.scheduler.client.report [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Deleted allocations for instance 2ac801d7-af70-46e0-88b3-02caee13497d [ 1012.823141] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1012.823141] env[62558]: value = "task-1267254" [ 1012.823141] env[62558]: _type = "Task" [ 1012.823141] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.839153] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267254, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.954224] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.963929] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267253, 'name': ReconfigVM_Task, 'duration_secs': 0.404994} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.964407] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1013.277341] env[62558]: INFO nova.compute.manager [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Took 13.60 seconds to build instance. [ 1013.337301] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8ab64de7-3134-43ae-a876-30f268a4e9cc tempest-ServerRescueNegativeTestJSON-687222775 tempest-ServerRescueNegativeTestJSON-687222775-project-member] Lock "2ac801d7-af70-46e0-88b3-02caee13497d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.497s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.343995] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267254, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.465018] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa62a217-cf61-4c6b-a6b1-1bade18b3f1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.473399] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.473817] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.474117] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.474408] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.474706] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.474970] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.475292] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.475571] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.475860] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.476157] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.476439] env[62558]: DEBUG nova.virt.hardware [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.481795] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1013.484013] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05ee163e-ef41-4f45-9f20-c323fa831066 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.497522] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85050beb-6240-4682-b241-8905a6e0373a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.533853] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a5425f-cbcd-4311-87a0-f17c09e869b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.537512] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1013.537512] env[62558]: value = "task-1267255" [ 1013.537512] env[62558]: _type = "Task" [ 1013.537512] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.546279] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6e84ba-c8c9-4d64-869b-883bb50cfb6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.555196] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267255, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.565351] env[62558]: DEBUG nova.compute.provider_tree [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.782105] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b87a3014-af77-4502-a4dc-00d80bcd91e9 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.105s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.841685] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267254, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575482} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.841956] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.842198] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.842482] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2839c54-72a3-427c-8cb2-7b93f7543bdb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.848881] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1013.848881] env[62558]: value = "task-1267256" [ 1013.848881] env[62558]: _type = "Task" [ 1013.848881] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.857152] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267256, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.053114] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267255, 'name': ReconfigVM_Task, 'duration_secs': 0.156948} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.053730] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1014.055534] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7864a1f3-a3b6-46b6-b749-2d85751858b2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.076479] env[62558]: DEBUG nova.scheduler.client.report [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.087705] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.088735] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-991990bd-e2e6-4a9f-b5d9-194b2e53391d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.112179] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1014.112179] env[62558]: value = "task-1267257" [ 1014.112179] env[62558]: _type = "Task" [ 1014.112179] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.123694] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267257, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.360870] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267256, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071641} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.361345] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.362683] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bb3f66-1f08-4f8e-addf-c96b87215682 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.388544] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.388961] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e343b57-7f77-425d-b814-57178a1eb0d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.413587] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1014.413587] env[62558]: value = "task-1267258" [ 1014.413587] env[62558]: _type = "Task" [ 1014.413587] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.426832] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267258, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.589977] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.593537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.639s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.593880] env[62558]: DEBUG nova.objects.instance [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'resources' on Instance uuid 6687a008-3193-4a43-b67c-c10e5ffcfc11 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.628444] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267257, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.629633] env[62558]: INFO nova.scheduler.client.report [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted allocations for instance 29584082-37e6-4dd8-906f-fd87b4ca9bc2 [ 1014.669872] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1014.671034] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1014.922825] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "a94f43ba-da37-479c-b939-016dff83c754" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.923113] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.924093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "a94f43ba-da37-479c-b939-016dff83c754-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.924093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.924093] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.926825] env[62558]: INFO nova.compute.manager [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Terminating instance [ 1014.927350] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267258, 'name': ReconfigVM_Task, 'duration_secs': 0.287897} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.927598] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 109f67db-21cf-40f9-966e-4730bff98bcf/109f67db-21cf-40f9-966e-4730bff98bcf.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.928952] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'device_type': 'disk', 'encryption_format': None, 'boot_index': 0, 'encryption_secret_uuid': None, 'disk_bus': None, 'encrypted': False, 'guest_format': None, 'encryption_options': None, 'size': 0, 'image_id': 'bd09177c-5fd7-495c-893b-5ec219b6b3de'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '37cca107-aa9f-46bf-8d51-3a067b109b01', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'}, 'boot_index': None, 'disk_bus': None, 'delete_on_termination': False, 'guest_format': None, 'mount_device': '/dev/sdb', 'volume_type': None}], 'swap': None} {{(pid=62558) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1014.929176] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1014.929369] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1014.935017] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28980b9d-762f-4b7c-9de0-396d953364d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.936354] env[62558]: DEBUG nova.compute.manager [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1014.936551] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.937293] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8899c7b6-c351-417a-b3b2-738f64b58cc6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.961399] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19101a79-2c23-4b0a-b17f-5bcb4284b565 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.962729] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.962973] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b06c6ab0-22e6-48c5-a206-aaa5d62feb18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.986117] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.987840] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c26c2ea1-4a41-4c56-ad2f-e48717a87999 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.001264] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1015.001264] env[62558]: value = "task-1267259" [ 1015.001264] env[62558]: _type = "Task" [ 1015.001264] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.006927] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1015.006927] env[62558]: value = "task-1267260" [ 1015.006927] env[62558]: _type = "Task" [ 1015.006927] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.010409] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.017925] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.123295] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267257, 'name': ReconfigVM_Task, 'duration_secs': 0.512272} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.123575] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101/9b87ed06-b062-4e6f-9dfa-a9b5491fe101.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.123952] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.142245] env[62558]: DEBUG oslo_concurrency.lockutils [None req-35d670bd-7e07-4dab-a3a7-6f5f8085b85a tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "29584082-37e6-4dd8-906f-fd87b4ca9bc2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.064s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.277062] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daddd700-a17a-4253-ae64-ce5e53d3d2cd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.284702] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce65d88-d561-48c6-836d-c75b38859e72 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.319025] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad6a1cf-b78e-4075-8b8b-2d7d0e2cd7a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.324102] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b75de-3d86-4720-99a0-ef72439d669b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.340024] env[62558]: DEBUG nova.compute.provider_tree [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.513582] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267259, 'name': PowerOffVM_Task, 'duration_secs': 0.192434} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.517010] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.517241] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.517510] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fe6fed0-ea5b-4adc-87dc-ea7f78ee7494 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.523562] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267260, 'name': ReconfigVM_Task, 'duration_secs': 0.317856} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.523826] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.528498] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44de8f98-f9d6-49e0-82fc-8d0124014a1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.544707] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1015.544707] env[62558]: value = "task-1267262" [ 1015.544707] env[62558]: _type = "Task" [ 1015.544707] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.552875] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.579744] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.580034] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.580305] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleting the datastore file [datastore2] a94f43ba-da37-479c-b939-016dff83c754 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.580580] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d28b509-c641-4f96-8323-c0a18c85b717 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.587567] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for the task: (returnval){ [ 1015.587567] env[62558]: value = "task-1267263" [ 1015.587567] env[62558]: _type = "Task" [ 1015.587567] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.597816] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267263, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.632910] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16852741-7117-40ff-937d-cbcd8884065e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.655322] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb9204a-cca5-4982-add7-aad6f8facbda {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.677874] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1015.842659] env[62558]: DEBUG nova.scheduler.client.report [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.055383] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267262, 'name': ReconfigVM_Task, 'duration_secs': 0.154004} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.055481] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1016.056603] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0e1b1dc-982c-4c57-90d3-991a66992b03 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.064229] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1016.064229] env[62558]: value = "task-1267264" [ 1016.064229] env[62558]: _type = "Task" [ 1016.064229] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.073899] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267264, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.097803] env[62558]: DEBUG oslo_vmware.api [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Task: {'id': task-1267263, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165264} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.098089] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.098284] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.098462] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.098638] env[62558]: INFO nova.compute.manager [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] [instance: a94f43ba-da37-479c-b939-016dff83c754] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1016.098890] env[62558]: DEBUG oslo.service.loopingcall [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.099104] env[62558]: DEBUG nova.compute.manager [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.099200] env[62558]: DEBUG nova.network.neutron [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.347796] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.381493] env[62558]: INFO nova.scheduler.client.report [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance 6687a008-3193-4a43-b67c-c10e5ffcfc11 [ 1016.416029] env[62558]: DEBUG nova.network.neutron [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Port 5bb4afdb-999e-446a-ab6e-1bd7a8bc092b binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1016.575780] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267264, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.580096] env[62558]: DEBUG nova.compute.manager [req-67dc5627-c5d7-42b9-a994-574c49424243 req-9896421c-9546-40a4-8f9d-74acdd0ffea2 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Received event network-vif-deleted-236df18f-ee35-44b7-80eb-4bb9408b4ae2 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.580096] env[62558]: INFO nova.compute.manager [req-67dc5627-c5d7-42b9-a994-574c49424243 req-9896421c-9546-40a4-8f9d-74acdd0ffea2 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Neutron deleted interface 236df18f-ee35-44b7-80eb-4bb9408b4ae2; detaching it from the instance and deleting it from the info cache [ 1016.580096] env[62558]: DEBUG nova.network.neutron [req-67dc5627-c5d7-42b9-a994-574c49424243 req-9896421c-9546-40a4-8f9d-74acdd0ffea2 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.891091] env[62558]: DEBUG oslo_concurrency.lockutils [None req-72384b8b-1e40-4ecb-888b-cbc2fa17c1bd tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "6687a008-3193-4a43-b67c-c10e5ffcfc11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.266s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.049912] env[62558]: DEBUG nova.network.neutron [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.076198] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267264, 'name': Rename_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.086680] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-743dd4c8-00e9-400f-8d03-edd5a0cf9b17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.096477] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b291989f-9885-4e6a-8b69-f4930d4e6d95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.125901] env[62558]: DEBUG nova.compute.manager [req-67dc5627-c5d7-42b9-a994-574c49424243 req-9896421c-9546-40a4-8f9d-74acdd0ffea2 service nova] [instance: a94f43ba-da37-479c-b939-016dff83c754] Detach interface failed, port_id=236df18f-ee35-44b7-80eb-4bb9408b4ae2, reason: Instance a94f43ba-da37-479c-b939-016dff83c754 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1017.459749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.459749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.459749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.553037] env[62558]: INFO nova.compute.manager [-] [instance: a94f43ba-da37-479c-b939-016dff83c754] Took 1.45 seconds to deallocate network for instance. [ 1017.575752] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267264, 'name': Rename_Task, 'duration_secs': 1.156963} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.577794] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.578110] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32051e99-f3f6-4ea3-9d2a-2ffbad18d829 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.588049] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1017.588049] env[62558]: value = "task-1267265" [ 1017.588049] env[62558]: _type = "Task" [ 1017.588049] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.601403] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267265, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.063650] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.066078] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.066078] env[62558]: DEBUG nova.objects.instance [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lazy-loading 'resources' on Instance uuid a94f43ba-da37-479c-b939-016dff83c754 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.091635] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.091904] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.092911] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "4dffea0b-adb7-4c6a-b2b0-436796111958" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.093140] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.105031] env[62558]: DEBUG oslo_vmware.api [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267265, 'name': PowerOnVM_Task, 'duration_secs': 0.457677} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.105621] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.105894] env[62558]: DEBUG nova.compute.manager [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.107183] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b9e5c7-3d33-41d9-adb7-41efd4640485 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.505454] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.505592] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.505707] env[62558]: DEBUG nova.network.neutron [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.598166] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.600834] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.629995] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.729251] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ff75e3-99d6-4957-af85-9ee603079df3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.735567] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056624df-2317-4a09-b22e-2cdacd1fc3b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.765883] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac28a9ee-ed49-47d8-a491-91ce59b4ae38 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.773357] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1325e033-4f63-44c5-90f1-4fa8a9226886 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.786825] env[62558]: DEBUG nova.compute.provider_tree [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.022751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "e7569632-751d-4bd7-b00a-5176c816551c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.022751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.127640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.128614] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.290453] env[62558]: DEBUG nova.scheduler.client.report [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.313679] env[62558]: DEBUG nova.network.neutron [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.526809] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1019.796456] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.800780] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.168s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.800780] env[62558]: DEBUG nova.objects.instance [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1019.816983] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.823684] env[62558]: INFO nova.scheduler.client.report [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Deleted allocations for instance a94f43ba-da37-479c-b939-016dff83c754 [ 1020.048284] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.326617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.326949] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.332892] env[62558]: DEBUG oslo_concurrency.lockutils [None req-ca8a9c63-91b9-4436-8fdd-c33a773df281 tempest-ServerDiskConfigTestJSON-1813885824 tempest-ServerDiskConfigTestJSON-1813885824-project-member] Lock "a94f43ba-da37-479c-b939-016dff83c754" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.409s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.342991] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42febe0a-7fe2-4cf7-aaf9-d39c2160ba4b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.361975] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9291d3e4-9493-4c1b-863a-0659dd7ce834 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.368887] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1020.808197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-2d25f91d-33b2-43b3-aeb4-9e99360f7d22 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.809476] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.682s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.810899] env[62558]: INFO nova.compute.claims [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.833211] env[62558]: INFO nova.compute.manager [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Detaching volume d479af90-d34c-43cd-a255-317c4b2c3d13 [ 1020.874688] env[62558]: INFO nova.virt.block_device [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Attempting to driver detach volume d479af90-d34c-43cd-a255-317c4b2c3d13 from mountpoint /dev/sdb [ 1020.874688] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1020.874835] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1020.877776] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.877776] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200eb734-f5b7-40e0-85fb-b815bc9c9aeb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.880357] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-389272fd-6cf5-4133-af1c-66ab4b4cef1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.902502] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d57359d-493b-434d-a7e6-c2fe935c57e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.905193] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1020.905193] env[62558]: value = "task-1267266" [ 1020.905193] env[62558]: _type = "Task" [ 1020.905193] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.911789] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5052db-3a88-4770-b2e0-2c2867f55d71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.917223] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267266, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.940016] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8371448-2091-4b9d-b37d-e648d2f0686c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.954970] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] The volume has not been displaced from its original location: [datastore2] volume-d479af90-d34c-43cd-a255-317c4b2c3d13/volume-d479af90-d34c-43cd-a255-317c4b2c3d13.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1020.960255] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1020.960601] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89b47aa7-d35a-499d-8f8a-b16b52b7a7f1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.978460] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1020.978460] env[62558]: value = "task-1267267" [ 1020.978460] env[62558]: _type = "Task" [ 1020.978460] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.986334] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267267, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.415827] env[62558]: DEBUG oslo_vmware.api [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267266, 'name': PowerOnVM_Task, 'duration_secs': 0.507747} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.416665] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.416935] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9b475432-ffc4-4d24-8d4a-cead812e81dd tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance '9b87ed06-b062-4e6f-9dfa-a9b5491fe101' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.491185] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267267, 'name': ReconfigVM_Task, 'duration_secs': 0.203681} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.492480] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1021.497753] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8073b5e5-689a-4bc8-b45e-e6afabee9b6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.515326] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1021.515326] env[62558]: value = "task-1267268" [ 1021.515326] env[62558]: _type = "Task" [ 1021.515326] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.524093] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267268, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.966627] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8677fdf-c310-45b6-a4f3-f346566e2b5a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.974565] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e95210-c8a3-4614-86e2-5cecfc391454 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.011691] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11a48b3-a9a8-4453-8566-fa8060edb591 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.023000] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883e96e2-ed0c-444b-af8f-cbdd85ef3a86 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.030594] env[62558]: DEBUG oslo_vmware.api [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267268, 'name': ReconfigVM_Task, 'duration_secs': 0.138776} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.031323] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272606', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'name': 'volume-d479af90-d34c-43cd-a255-317c4b2c3d13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '109f67db-21cf-40f9-966e-4730bff98bcf', 'attached_at': '', 'detached_at': '', 'volume_id': 'd479af90-d34c-43cd-a255-317c4b2c3d13', 'serial': 'd479af90-d34c-43cd-a255-317c4b2c3d13'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1022.041860] env[62558]: DEBUG nova.compute.provider_tree [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.546093] env[62558]: DEBUG nova.scheduler.client.report [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.577911] env[62558]: DEBUG nova.objects.instance [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'flavor' on Instance uuid 109f67db-21cf-40f9-966e-4730bff98bcf {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.049935] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.240s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.050549] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.053072] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.924s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.055035] env[62558]: INFO nova.compute.claims [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.559111] env[62558]: DEBUG nova.compute.utils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.562831] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1023.563010] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.584857] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7a7f8ce2-846c-413c-86c5-e85090777f9a tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.258s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.610891] env[62558]: DEBUG nova.policy [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1684539271b4820b0f6f53b3b44898b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c184479dcbc849ea983347809d5fc3b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1023.869431] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.869749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.869956] env[62558]: DEBUG nova.compute.manager [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Going to confirm migration 3 {{(pid=62558) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1023.908113] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Successfully created port: 68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.063902] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1024.230073] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e984776f-d80d-4e55-8959-55cea804ff37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.238999] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8cd18a-82d7-47a9-8d8f-e0cb882f6c65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.270025] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6d6fd0-19e6-449d-807a-9c53247f12a0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.279557] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd1a968-d405-4f8b-b019-7c2f75634c55 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.294490] env[62558]: DEBUG nova.compute.provider_tree [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.432636] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.432865] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquired lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.433060] env[62558]: DEBUG nova.network.neutron [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.433256] env[62558]: DEBUG nova.objects.instance [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'info_cache' on Instance uuid 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.640290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.640661] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.640759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.640946] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.641159] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.643168] env[62558]: INFO nova.compute.manager [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Terminating instance [ 1024.645040] env[62558]: DEBUG nova.compute.manager [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.645242] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.646159] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4d6b59-7d35-4000-9fbc-5014eeab4c99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.653968] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.654216] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb7aad81-eff1-4e65-8172-c3019745e4ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.660387] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1024.660387] env[62558]: value = "task-1267269" [ 1024.660387] env[62558]: _type = "Task" [ 1024.660387] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.668814] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.797724] env[62558]: DEBUG nova.scheduler.client.report [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.074979] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1025.103923] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.104202] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.104364] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.104546] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.104695] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.104960] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.105226] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.105406] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.105577] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.105742] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.106031] env[62558]: DEBUG nova.virt.hardware [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.106875] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf870e4c-37fa-43d0-9d0f-52b12e3728ac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.114397] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1730e8f9-869b-4296-b8a4-0ee66ef6507c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.169654] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267269, 'name': PowerOffVM_Task, 'duration_secs': 0.24052} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.169936] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.170156] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.170405] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c04b648f-8dac-49c7-a686-b7dd5cb5c20c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.265080] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.265293] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.265480] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore1] 109f67db-21cf-40f9-966e-4730bff98bcf {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.265746] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e686fe7-3475-4f71-8b67-8b699063419d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.273136] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1025.273136] env[62558]: value = "task-1267271" [ 1025.273136] env[62558]: _type = "Task" [ 1025.273136] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.281611] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.302580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.303230] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.306087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.258s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.307555] env[62558]: INFO nova.compute.claims [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.458120] env[62558]: DEBUG nova.compute.manager [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Received event network-vif-plugged-68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.458426] env[62558]: DEBUG oslo_concurrency.lockutils [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] Acquiring lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.458702] env[62558]: DEBUG oslo_concurrency.lockutils [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.458914] env[62558]: DEBUG oslo_concurrency.lockutils [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.459155] env[62558]: DEBUG nova.compute.manager [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] No waiting events found dispatching network-vif-plugged-68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.459387] env[62558]: WARNING nova.compute.manager [req-1903e6a5-cc0c-40d0-9a7a-5603bb703e6e req-cd06f829-071d-4dba-8fce-1abdbedf806d service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Received unexpected event network-vif-plugged-68f78b29-4c73-4a2c-b92f-1d34245b9c68 for instance with vm_state building and task_state spawning. [ 1025.551253] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Successfully updated port: 68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.631034] env[62558]: DEBUG nova.network.neutron [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [{"id": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "address": "fa:16:3e:16:6b:63", "network": {"id": "1c3643ee-8eac-4018-834e-c738657200d1", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1701187674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a090a1a385e43dc840b63eabb74d0cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f78b07ea-f425-4622-84f4-706a5d8820a7", "external-id": "nsx-vlan-transportzone-126", "segmentation_id": 126, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb4afdb-99", "ovs_interfaceid": "5bb4afdb-999e-446a-ab6e-1bd7a8bc092b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.782819] env[62558]: DEBUG oslo_vmware.api [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215752} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.783192] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.783240] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.783408] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.783586] env[62558]: INFO nova.compute.manager [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1025.783826] env[62558]: DEBUG oslo.service.loopingcall [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.784034] env[62558]: DEBUG nova.compute.manager [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.784131] env[62558]: DEBUG nova.network.neutron [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1025.812121] env[62558]: DEBUG nova.compute.utils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.815664] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.815812] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.961976] env[62558]: DEBUG nova.policy [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.053751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.053905] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.054087] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.133630] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Releasing lock "refresh_cache-9b87ed06-b062-4e6f-9dfa-a9b5491fe101" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.134079] env[62558]: DEBUG nova.objects.instance [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lazy-loading 'migration_context' on Instance uuid 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.281060] env[62558]: DEBUG nova.compute.manager [req-cd1248f8-d46a-478c-ba89-b685c92d97ed req-6df5f9f2-9fe8-4590-be9b-a32ff83ea7b4 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Received event network-vif-deleted-94d89474-fbcc-4208-aa16-91863dd3e2e5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.281279] env[62558]: INFO nova.compute.manager [req-cd1248f8-d46a-478c-ba89-b685c92d97ed req-6df5f9f2-9fe8-4590-be9b-a32ff83ea7b4 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Neutron deleted interface 94d89474-fbcc-4208-aa16-91863dd3e2e5; detaching it from the instance and deleting it from the info cache [ 1026.281460] env[62558]: DEBUG nova.network.neutron [req-cd1248f8-d46a-478c-ba89-b685c92d97ed req-6df5f9f2-9fe8-4590-be9b-a32ff83ea7b4 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.318675] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.321569] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Successfully created port: dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.456830] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9599bc9-d978-42d9-9b96-d867d06ae31a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.472227] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1666e9-6030-4309-bd98-0f9874f758ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.511471] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc9bcc0-5ccc-40a9-9601-6b2190290ed4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.519541] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2f66c2-9671-4c80-b8a0-227a8dce9dbd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.534985] env[62558]: DEBUG nova.compute.provider_tree [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.586855] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1026.637659] env[62558]: DEBUG nova.objects.base [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Object Instance<9b87ed06-b062-4e6f-9dfa-a9b5491fe101> lazy-loaded attributes: info_cache,migration_context {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1026.638606] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eea525b-4941-4b79-b6c9-59c0e18212e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.664973] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfd241f6-9885-46f7-9015-3d4f23d16dfa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.670887] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1026.670887] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52607b12-4792-c868-de26-83406fd13837" [ 1026.670887] env[62558]: _type = "Task" [ 1026.670887] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.678828] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52607b12-4792-c868-de26-83406fd13837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.723513] env[62558]: DEBUG nova.network.neutron [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.730898] env[62558]: DEBUG nova.network.neutron [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Updating instance_info_cache with network_info: [{"id": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "address": "fa:16:3e:f2:04:73", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68f78b29-4c", "ovs_interfaceid": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.784939] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac4b1089-1547-45c5-b04a-1676c92382d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.794439] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1b6f9c-b677-42da-b744-11a37e1b9c37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.820580] env[62558]: DEBUG nova.compute.manager [req-cd1248f8-d46a-478c-ba89-b685c92d97ed req-6df5f9f2-9fe8-4590-be9b-a32ff83ea7b4 service nova] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Detach interface failed, port_id=94d89474-fbcc-4208-aa16-91863dd3e2e5, reason: Instance 109f67db-21cf-40f9-966e-4730bff98bcf could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.038101] env[62558]: DEBUG nova.scheduler.client.report [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.180933] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52607b12-4792-c868-de26-83406fd13837, 'name': SearchDatastore_Task, 'duration_secs': 0.006833} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.181251] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.226344] env[62558]: INFO nova.compute.manager [-] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Took 1.44 seconds to deallocate network for instance. [ 1027.233834] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.233834] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Instance network_info: |[{"id": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "address": "fa:16:3e:f2:04:73", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68f78b29-4c", "ovs_interfaceid": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1027.234185] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:04:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86b8f7fc-c105-4bcb-a4ec-c363ed38b17a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '68f78b29-4c73-4a2c-b92f-1d34245b9c68', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.241699] env[62558]: DEBUG oslo.service.loopingcall [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.241905] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.242142] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2390946-b700-4d7a-97d4-4092fdd183c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.264595] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.264595] env[62558]: value = "task-1267272" [ 1027.264595] env[62558]: _type = "Task" [ 1027.264595] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.272794] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267272, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.332960] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.362694] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.363045] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.363226] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.363416] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.363565] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.363718] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.363929] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.364106] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.364279] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.364444] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.364619] env[62558]: DEBUG nova.virt.hardware [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.365553] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f06ad7-6ddc-4983-982e-d4f15e634f33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.374764] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de877ee-06b1-4b3c-81a9-9de82ff27370 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.486235] env[62558]: DEBUG nova.compute.manager [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Received event network-changed-68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.486536] env[62558]: DEBUG nova.compute.manager [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Refreshing instance network info cache due to event network-changed-68f78b29-4c73-4a2c-b92f-1d34245b9c68. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1027.486886] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] Acquiring lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.487141] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] Acquired lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.487372] env[62558]: DEBUG nova.network.neutron [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Refreshing network info cache for port 68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1027.543079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.543635] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1027.546568] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.365s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.732415] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.775231] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267272, 'name': CreateVM_Task, 'duration_secs': 0.335016} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.775231] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1027.776017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.776215] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.776539] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.776801] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7dd6881-f1aa-4fc0-ba63-3c4bdbcfc67f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.781539] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1027.781539] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f4c4d8-d52f-3726-84e1-3c1e5783003f" [ 1027.781539] env[62558]: _type = "Task" [ 1027.781539] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.789431] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f4c4d8-d52f-3726-84e1-3c1e5783003f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.805669] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Successfully updated port: dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.049745] env[62558]: DEBUG nova.compute.utils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1028.053898] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1028.054132] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1028.096061] env[62558]: DEBUG nova.policy [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15d30a8abc2e415797ae2ee317b9d7e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '565ee1048ca449cf8755dce51cb4b0b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1028.201759] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79a3301-5e20-445b-a40b-bb9bb7999913 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.210052] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2364d6-ce9c-4d5c-811c-50e9eb158fcc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.244377] env[62558]: DEBUG nova.network.neutron [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Updated VIF entry in instance network info cache for port 68f78b29-4c73-4a2c-b92f-1d34245b9c68. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.244719] env[62558]: DEBUG nova.network.neutron [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Updating instance_info_cache with network_info: [{"id": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "address": "fa:16:3e:f2:04:73", "network": {"id": "4ae32a58-b56b-48db-b34c-8cc391d0a9b3", "bridge": "br-int", "label": "tempest-ServersTestJSON-1937124211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c184479dcbc849ea983347809d5fc3b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86b8f7fc-c105-4bcb-a4ec-c363ed38b17a", "external-id": "nsx-vlan-transportzone-830", "segmentation_id": 830, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68f78b29-4c", "ovs_interfaceid": "68f78b29-4c73-4a2c-b92f-1d34245b9c68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.246380] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2e3516-40ee-440f-a3d1-82c389e1700e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.254193] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2095a1ba-591e-4663-9c12-266b6cb0ea78 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.268265] env[62558]: DEBUG nova.compute.provider_tree [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.291479] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f4c4d8-d52f-3726-84e1-3c1e5783003f, 'name': SearchDatastore_Task, 'duration_secs': 0.046873} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.291786] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.292030] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1028.292273] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.292420] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.292596] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1028.292852] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-459507a3-ce09-4dca-9bdb-90fab22240b1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.307605] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.307797] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1028.309486] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47d68874-397c-468d-baac-1f5b934d6b64 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.311034] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.311034] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.311212] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.315471] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1028.315471] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f44a94-c419-c086-d962-d19a54121cee" [ 1028.315471] env[62558]: _type = "Task" [ 1028.315471] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.322958] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f44a94-c419-c086-d962-d19a54121cee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.346306] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Successfully created port: 2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.554617] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1028.749958] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d30ca49-8f79-4a27-8f55-af202f841e10 req-dcf711f5-dc6f-4b87-a800-ed8d3e5b560a service nova] Releasing lock "refresh_cache-4dffea0b-adb7-4c6a-b2b0-436796111958" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.771335] env[62558]: DEBUG nova.scheduler.client.report [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.825670] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f44a94-c419-c086-d962-d19a54121cee, 'name': SearchDatastore_Task, 'duration_secs': 0.030851} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.826525] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a151140f-01ad-4264-850f-4af9d040f55a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.831791] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1028.831791] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52029cb2-50ec-3ef0-0258-33e3ec5f4ffd" [ 1028.831791] env[62558]: _type = "Task" [ 1028.831791] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.839242] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52029cb2-50ec-3ef0-0258-33e3ec5f4ffd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.845975] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.978233] env[62558]: DEBUG nova.network.neutron [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.259774] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.260105] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.343265] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52029cb2-50ec-3ef0-0258-33e3ec5f4ffd, 'name': SearchDatastore_Task, 'duration_secs': 0.010998} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.343515] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.343771] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 4dffea0b-adb7-4c6a-b2b0-436796111958/4dffea0b-adb7-4c6a-b2b0-436796111958.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.344031] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-afb56006-7fbd-4343-b6c1-03723ab74281 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.350100] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1029.350100] env[62558]: value = "task-1267273" [ 1029.350100] env[62558]: _type = "Task" [ 1029.350100] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.357787] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267273, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.480790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.481270] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Instance network_info: |[{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.481830] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:7e:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dedfab3d-cdf6-40b7-953a-7d13dad8f5a6', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.490345] env[62558]: DEBUG oslo.service.loopingcall [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.490674] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.490958] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a41565bf-04b3-405e-997c-f356aa76e525 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.512647] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.512647] env[62558]: value = "task-1267274" [ 1029.512647] env[62558]: _type = "Task" [ 1029.512647] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.521492] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267274, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.564504] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1029.595130] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.595515] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.595736] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.595996] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.596165] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.596329] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.596609] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.596780] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.596954] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.597186] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.597375] env[62558]: DEBUG nova.virt.hardware [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.598389] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c994e1fc-f955-4047-ac24-aee936c982e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.607270] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04514bc6-0411-444e-9492-aa8fe87c7942 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.764757] env[62558]: DEBUG nova.compute.utils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.780784] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.234s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.783807] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.051s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.784019] env[62558]: DEBUG nova.objects.instance [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'resources' on Instance uuid 109f67db-21cf-40f9-966e-4730bff98bcf {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.825375] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Successfully updated port: 2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.859646] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267273, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442454} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.859952] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 4dffea0b-adb7-4c6a-b2b0-436796111958/4dffea0b-adb7-4c6a-b2b0-436796111958.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.860111] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.860374] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e5aa2a1-33ba-49cf-8b61-4c931aa7cfd7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.866313] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1029.866313] env[62558]: value = "task-1267275" [ 1029.866313] env[62558]: _type = "Task" [ 1029.866313] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.874019] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267275, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.023352] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267274, 'name': CreateVM_Task, 'duration_secs': 0.412091} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.023611] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.024589] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.024814] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.025300] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.025570] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf64caf-4b08-41c7-81f9-cc03ecd54be8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.030174] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1030.030174] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52edca78-0ada-7630-c7b0-14540e4dc0be" [ 1030.030174] env[62558]: _type = "Task" [ 1030.030174] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.037893] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52edca78-0ada-7630-c7b0-14540e4dc0be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.268589] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.329737] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.329919] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquired lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.330040] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.341110] env[62558]: INFO nova.scheduler.client.report [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocation for migration 24d51b19-2879-4b7f-a7de-4e4f1b53d301 [ 1030.376427] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267275, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073686} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.376694] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.377917] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737ae3fd-79d1-42f5-a15b-6d84919b1d60 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.403622] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 4dffea0b-adb7-4c6a-b2b0-436796111958/4dffea0b-adb7-4c6a-b2b0-436796111958.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.406404] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5709d60f-143d-46cf-8c46-050610c606e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.425792] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1030.425792] env[62558]: value = "task-1267276" [ 1030.425792] env[62558]: _type = "Task" [ 1030.425792] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.434447] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.441142] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fc653b-050e-4a6e-8340-76cb6e9cf7e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.448133] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84709468-5e9b-43e9-8634-9a01bde21589 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.477673] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ba29a6-ee7c-424c-be5e-4cdb1a4148b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.484527] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfce060-330e-4f95-b928-98d0e4840bd2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.497288] env[62558]: DEBUG nova.compute.provider_tree [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.539912] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52edca78-0ada-7630-c7b0-14540e4dc0be, 'name': SearchDatastore_Task, 'duration_secs': 0.031464} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.540255] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.540493] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.540732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.540881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.541082] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.541346] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f987ae31-a88b-4c2c-a2af-d19a8202526d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.549163] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.549351] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.550065] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07c1c34b-a1f4-4ab2-a46b-9e04515b8139 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.554754] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1030.554754] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523cc141-0a10-68d0-2eb0-80eb205c400a" [ 1030.554754] env[62558]: _type = "Task" [ 1030.554754] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.563809] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523cc141-0a10-68d0-2eb0-80eb205c400a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.847991] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.978s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.863911] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.935467] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267276, 'name': ReconfigVM_Task, 'duration_secs': 0.272737} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.935752] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 4dffea0b-adb7-4c6a-b2b0-436796111958/4dffea0b-adb7-4c6a-b2b0-436796111958.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.936446] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8aadda9-6a51-45ea-8cc9-3cb962b3acf3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.942824] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1030.942824] env[62558]: value = "task-1267277" [ 1030.942824] env[62558]: _type = "Task" [ 1030.942824] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.950513] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267277, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.991330] env[62558]: DEBUG nova.network.neutron [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Updating instance_info_cache with network_info: [{"id": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "address": "fa:16:3e:6a:aa:6f", "network": {"id": "c2bf894a-1908-43cc-8b72-5715f4ab70a6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-488993050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "565ee1048ca449cf8755dce51cb4b0b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74c816b2-b8b0-432e-baac-662ed8ea0417", "external-id": "nsx-vlan-transportzone-776", "segmentation_id": 776, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e27a73e-58", "ovs_interfaceid": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.000392] env[62558]: DEBUG nova.scheduler.client.report [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1031.064949] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]523cc141-0a10-68d0-2eb0-80eb205c400a, 'name': SearchDatastore_Task, 'duration_secs': 0.009532} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.065789] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9df67281-a113-4109-aaca-ade26eb98b33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.070830] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1031.070830] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527fb4f0-0380-7265-0155-ae4d2ad068a7" [ 1031.070830] env[62558]: _type = "Task" [ 1031.070830] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.076950] env[62558]: DEBUG nova.compute.manager [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-vif-plugged-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.077248] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.077390] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.077553] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.077719] env[62558]: DEBUG nova.compute.manager [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] No waiting events found dispatching network-vif-plugged-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.077883] env[62558]: WARNING nova.compute.manager [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received unexpected event network-vif-plugged-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 for instance with vm_state building and task_state spawning. [ 1031.078055] env[62558]: DEBUG nova.compute.manager [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.078216] env[62558]: DEBUG nova.compute.manager [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.078402] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.078564] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.078725] env[62558]: DEBUG nova.network.neutron [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.082920] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527fb4f0-0380-7265-0155-ae4d2ad068a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.148173] env[62558]: DEBUG nova.compute.manager [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Received event network-vif-plugged-2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.148173] env[62558]: DEBUG oslo_concurrency.lockutils [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] Acquiring lock "e7569632-751d-4bd7-b00a-5176c816551c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.148173] env[62558]: DEBUG oslo_concurrency.lockutils [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] Lock "e7569632-751d-4bd7-b00a-5176c816551c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.148173] env[62558]: DEBUG oslo_concurrency.lockutils [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] Lock "e7569632-751d-4bd7-b00a-5176c816551c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.148310] env[62558]: DEBUG nova.compute.manager [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] No waiting events found dispatching network-vif-plugged-2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.148464] env[62558]: WARNING nova.compute.manager [req-91df33d2-e687-4c45-9bb5-2b4d1b18c0f1 req-39ede6c7-28e1-416a-8bd8-41b1d86f7323 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Received unexpected event network-vif-plugged-2e27a73e-589c-4e74-bfc4-23d8a521c175 for instance with vm_state building and task_state spawning. [ 1031.328678] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.328893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.329174] env[62558]: INFO nova.compute.manager [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attaching volume cee3b8af-c539-4269-8100-6e3f68136336 to /dev/sdb [ 1031.362308] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d61f4dc-dc03-4431-815d-980b17153401 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.369303] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedc9be8-80d8-4324-8bce-77cd692867ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.381692] env[62558]: DEBUG nova.virt.block_device [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating existing volume attachment record: 4ac70ba3-6d1c-4711-b06e-2cc7eb6c741e {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1031.452226] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267277, 'name': Rename_Task, 'duration_secs': 0.139403} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.452529] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.452808] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-caca163b-74d7-4beb-ab42-5129b2659fb0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.459807] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1031.459807] env[62558]: value = "task-1267278" [ 1031.459807] env[62558]: _type = "Task" [ 1031.459807] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.468107] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.495071] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Releasing lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.496041] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Instance network_info: |[{"id": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "address": "fa:16:3e:6a:aa:6f", "network": {"id": "c2bf894a-1908-43cc-8b72-5715f4ab70a6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-488993050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "565ee1048ca449cf8755dce51cb4b0b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74c816b2-b8b0-432e-baac-662ed8ea0417", "external-id": "nsx-vlan-transportzone-776", "segmentation_id": 776, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e27a73e-58", "ovs_interfaceid": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.496041] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:aa:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74c816b2-b8b0-432e-baac-662ed8ea0417', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e27a73e-589c-4e74-bfc4-23d8a521c175', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.503627] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Creating folder: Project (565ee1048ca449cf8755dce51cb4b0b0). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1031.503911] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7fb1982-1719-4660-a371-d0e27087ca12 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.506199] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.516263] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Created folder: Project (565ee1048ca449cf8755dce51cb4b0b0) in parent group-v272451. [ 1031.516453] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Creating folder: Instances. Parent ref: group-v272612. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1031.516702] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b94c7803-bdd6-4deb-b083-32f4960846ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.523783] env[62558]: INFO nova.scheduler.client.report [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocations for instance 109f67db-21cf-40f9-966e-4730bff98bcf [ 1031.526051] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Created folder: Instances in parent group-v272612. [ 1031.526349] env[62558]: DEBUG oslo.service.loopingcall [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.528925] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.529271] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f61d6ca-3bc5-424f-97e2-30056cbc528e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.548670] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.548670] env[62558]: value = "task-1267282" [ 1031.548670] env[62558]: _type = "Task" [ 1031.548670] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.556133] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267282, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.579827] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527fb4f0-0380-7265-0155-ae4d2ad068a7, 'name': SearchDatastore_Task, 'duration_secs': 0.009664} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.580074] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.580352] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 5f75bf07-002f-4ed4-930a-fbe6f939cb4f/5f75bf07-002f-4ed4-930a-fbe6f939cb4f.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.580616] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-382fe3a4-6635-4c7a-a383-0ca00ee01f6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.591197] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1031.591197] env[62558]: value = "task-1267283" [ 1031.591197] env[62558]: _type = "Task" [ 1031.591197] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.600037] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.810463] env[62558]: DEBUG nova.network.neutron [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.810989] env[62558]: DEBUG nova.network.neutron [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.893150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.893573] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.893713] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.893915] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.894195] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.896899] env[62558]: INFO nova.compute.manager [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Terminating instance [ 1031.899348] env[62558]: DEBUG nova.compute.manager [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.899517] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.900391] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4cf3e7-c370-43f0-bb64-52c784b7d7e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.908574] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.908914] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc3b1507-bdca-4b80-a701-23293a2c22c8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.915137] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1031.915137] env[62558]: value = "task-1267284" [ 1031.915137] env[62558]: _type = "Task" [ 1031.915137] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.923205] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.970269] env[62558]: DEBUG oslo_vmware.api [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267278, 'name': PowerOnVM_Task, 'duration_secs': 0.445448} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.970557] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.970766] env[62558]: INFO nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Took 6.90 seconds to spawn the instance on the hypervisor. [ 1031.970951] env[62558]: DEBUG nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.971799] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46233263-0838-49d3-96c9-215251ced744 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.032571] env[62558]: DEBUG oslo_concurrency.lockutils [None req-7bf9168d-58e2-4976-86cc-838b526469d3 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "109f67db-21cf-40f9-966e-4730bff98bcf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.392s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.060010] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267282, 'name': CreateVM_Task, 'duration_secs': 0.293559} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.060010] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1032.060688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.060881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.061238] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.061530] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb47f3e0-d736-4c8b-a636-b2b01c38f0f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.065989] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1032.065989] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad002d-ee5c-93ff-426d-0af086e88638" [ 1032.065989] env[62558]: _type = "Task" [ 1032.065989] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.073888] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad002d-ee5c-93ff-426d-0af086e88638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.101763] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475196} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.102207] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 5f75bf07-002f-4ed4-930a-fbe6f939cb4f/5f75bf07-002f-4ed4-930a-fbe6f939cb4f.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.102376] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.102580] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12e87112-6f6c-4a79-85df-c8b1451f6795 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.108865] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1032.108865] env[62558]: value = "task-1267285" [ 1032.108865] env[62558]: _type = "Task" [ 1032.108865] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.116511] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.313900] env[62558]: DEBUG oslo_concurrency.lockutils [req-48340e98-b435-48a4-8cd8-852a4f3fb909 req-a7c7155d-81cf-4921-ae24-d257d2312c7a service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.425993] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267284, 'name': PowerOffVM_Task, 'duration_secs': 0.232238} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.426306] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.426504] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.426879] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3eb0f156-f9a7-4083-860b-06ba0d42738c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.489051] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.489051] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.489694] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleting the datastore file [datastore2] 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.492161] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49e868b6-c157-42c9-bfa9-0df99e8d7de4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.496585] env[62558]: INFO nova.compute.manager [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Took 13.39 seconds to build instance. [ 1032.501988] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for the task: (returnval){ [ 1032.501988] env[62558]: value = "task-1267287" [ 1032.501988] env[62558]: _type = "Task" [ 1032.501988] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.510202] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.576366] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ad002d-ee5c-93ff-426d-0af086e88638, 'name': SearchDatastore_Task, 'duration_secs': 0.016912} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.576681] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.576920] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.578218] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.578218] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.578218] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.578218] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-754cd8bd-615b-49ef-9c0b-e56255af1e22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.585539] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.586067] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.586372] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8471b33c-5e8f-4ee8-832b-2ba831cff3a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.592046] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1032.592046] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f52eb-2129-c77e-a8b5-f5044cc798e4" [ 1032.592046] env[62558]: _type = "Task" [ 1032.592046] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.599016] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f52eb-2129-c77e-a8b5-f5044cc798e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.617709] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267285, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066646} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.617929] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.618759] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b91a4e9-8a86-40b5-a2b1-6a40165c1daa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.641282] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 5f75bf07-002f-4ed4-930a-fbe6f939cb4f/5f75bf07-002f-4ed4-930a-fbe6f939cb4f.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.641588] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b3d63d5-d7f1-4631-a74c-4a6904bacc3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.661487] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1032.661487] env[62558]: value = "task-1267288" [ 1032.661487] env[62558]: _type = "Task" [ 1032.661487] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.669716] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267288, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.999396] env[62558]: DEBUG oslo_concurrency.lockutils [None req-02dd555a-5d5a-4b7e-8ad2-d768746bf1b6 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.906s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.013171] env[62558]: DEBUG oslo_vmware.api [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Task: {'id': task-1267287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140515} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.013496] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.013696] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.013897] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.014121] env[62558]: INFO nova.compute.manager [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1033.014442] env[62558]: DEBUG oslo.service.loopingcall [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.014617] env[62558]: DEBUG nova.compute.manager [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.014719] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1033.031216] env[62558]: DEBUG oslo_concurrency.lockutils [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "4dffea0b-adb7-4c6a-b2b0-436796111958" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.031629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.031766] env[62558]: DEBUG nova.compute.manager [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.032633] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f269e02-46d5-42eb-8c4d-f0c87399a279 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.040287] env[62558]: DEBUG nova.compute.manager [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1033.040857] env[62558]: DEBUG nova.objects.instance [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'flavor' on Instance uuid 4dffea0b-adb7-4c6a-b2b0-436796111958 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.108882] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f52eb-2129-c77e-a8b5-f5044cc798e4, 'name': SearchDatastore_Task, 'duration_secs': 0.007937} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.110437] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4be78537-b8a2-457f-9354-d6fcbc36faae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.117443] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1033.117443] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aba070-6a3b-a502-db59-702cc13f7ef9" [ 1033.117443] env[62558]: _type = "Task" [ 1033.117443] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.126354] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aba070-6a3b-a502-db59-702cc13f7ef9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.173086] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267288, 'name': ReconfigVM_Task, 'duration_secs': 0.267397} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.173549] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 5f75bf07-002f-4ed4-930a-fbe6f939cb4f/5f75bf07-002f-4ed4-930a-fbe6f939cb4f.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.174319] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffeec86d-1162-4d69-9064-0b9717f31fe1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.181462] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1033.181462] env[62558]: value = "task-1267289" [ 1033.181462] env[62558]: _type = "Task" [ 1033.181462] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.189098] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267289, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.311420] env[62558]: DEBUG nova.compute.manager [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Received event network-changed-2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.312170] env[62558]: DEBUG nova.compute.manager [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Refreshing instance network info cache due to event network-changed-2e27a73e-589c-4e74-bfc4-23d8a521c175. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1033.312369] env[62558]: DEBUG oslo_concurrency.lockutils [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] Acquiring lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.312674] env[62558]: DEBUG oslo_concurrency.lockutils [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] Acquired lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.313027] env[62558]: DEBUG nova.network.neutron [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Refreshing network info cache for port 2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.545821] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.546947] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0f80d99-9356-4589-b35c-c875058dcce6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.553070] env[62558]: DEBUG oslo_vmware.api [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1033.553070] env[62558]: value = "task-1267290" [ 1033.553070] env[62558]: _type = "Task" [ 1033.553070] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.560771] env[62558]: DEBUG oslo_vmware.api [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.601544] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "ea680004-b94a-47f9-b516-2afb95c30a51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.601843] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.602179] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.602441] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.602655] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.605386] env[62558]: INFO nova.compute.manager [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Terminating instance [ 1033.607501] env[62558]: DEBUG nova.compute.manager [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1033.607735] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.608593] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0af5f5d-4e95-43a5-a415-d491acdcc1e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.616924] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.617244] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d58ba79-6525-4977-9508-977126b30aa5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.628093] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aba070-6a3b-a502-db59-702cc13f7ef9, 'name': SearchDatastore_Task, 'duration_secs': 0.039759} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.629903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.630362] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] e7569632-751d-4bd7-b00a-5176c816551c/e7569632-751d-4bd7-b00a-5176c816551c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.630765] env[62558]: DEBUG oslo_vmware.api [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1033.630765] env[62558]: value = "task-1267292" [ 1033.630765] env[62558]: _type = "Task" [ 1033.630765] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.630994] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0462ae0a-e6c9-4dae-8b8b-4c539316fe81 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.642668] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1033.642668] env[62558]: value = "task-1267293" [ 1033.642668] env[62558]: _type = "Task" [ 1033.642668] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.650582] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.692836] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267289, 'name': Rename_Task, 'duration_secs': 0.143751} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.693216] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.693538] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c66b554b-aed6-4560-a239-4cce5949eda4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.700461] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1033.700461] env[62558]: value = "task-1267294" [ 1033.700461] env[62558]: _type = "Task" [ 1033.700461] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.709704] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.788469] env[62558]: DEBUG nova.network.neutron [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.063219] env[62558]: DEBUG oslo_vmware.api [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267290, 'name': PowerOffVM_Task, 'duration_secs': 0.1944} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.063624] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.063720] env[62558]: DEBUG nova.compute.manager [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.064593] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d79f9a-8e1d-4cbf-934f-b4171a16616b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.077392] env[62558]: DEBUG nova.network.neutron [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Updated VIF entry in instance network info cache for port 2e27a73e-589c-4e74-bfc4-23d8a521c175. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1034.077868] env[62558]: DEBUG nova.network.neutron [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Updating instance_info_cache with network_info: [{"id": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "address": "fa:16:3e:6a:aa:6f", "network": {"id": "c2bf894a-1908-43cc-8b72-5715f4ab70a6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-488993050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "565ee1048ca449cf8755dce51cb4b0b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74c816b2-b8b0-432e-baac-662ed8ea0417", "external-id": "nsx-vlan-transportzone-776", "segmentation_id": 776, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e27a73e-58", "ovs_interfaceid": "2e27a73e-589c-4e74-bfc4-23d8a521c175", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.142663] env[62558]: DEBUG oslo_vmware.api [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267292, 'name': PowerOffVM_Task, 'duration_secs': 0.203959} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.142890] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.143115] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.143450] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6532f9e-caf4-4a1c-8edf-d10234fcc877 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.153928] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267293, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469385} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.154230] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] e7569632-751d-4bd7-b00a-5176c816551c/e7569632-751d-4bd7-b00a-5176c816551c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.154426] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.154675] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9913d395-44ce-496d-9fde-ac0599d9e886 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.160907] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1034.160907] env[62558]: value = "task-1267296" [ 1034.160907] env[62558]: _type = "Task" [ 1034.160907] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.209836] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.210134] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.215077] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267294, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.291283] env[62558]: INFO nova.compute.manager [-] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Took 1.28 seconds to deallocate network for instance. [ 1034.304842] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.305113] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.305316] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleting the datastore file [datastore1] ea680004-b94a-47f9-b516-2afb95c30a51 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.305579] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc3c4085-d56e-4487-ad9b-f2d8279cfb0b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.312831] env[62558]: DEBUG oslo_vmware.api [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1034.312831] env[62558]: value = "task-1267297" [ 1034.312831] env[62558]: _type = "Task" [ 1034.312831] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.321345] env[62558]: DEBUG oslo_vmware.api [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.579043] env[62558]: DEBUG oslo_concurrency.lockutils [None req-675dd410-563e-457b-801f-fc633eb32f6f tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.580968] env[62558]: DEBUG oslo_concurrency.lockutils [req-e8a3a211-c0ac-45ad-b9a8-1a158a03ce86 req-d480f271-f952-4622-9377-f674b17c65e3 service nova] Releasing lock "refresh_cache-e7569632-751d-4bd7-b00a-5176c816551c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.670995] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267296, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094333} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.671298] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.672110] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebc7164-7d6a-4f8d-b727-344cb39740ed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.693669] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e7569632-751d-4bd7-b00a-5176c816551c/e7569632-751d-4bd7-b00a-5176c816551c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.693953] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7310178f-bc05-43b0-a93d-23b509584289 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.717874] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1034.720417] env[62558]: DEBUG oslo_vmware.api [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267294, 'name': PowerOnVM_Task, 'duration_secs': 0.520224} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.721998] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.722239] env[62558]: INFO nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Took 7.39 seconds to spawn the instance on the hypervisor. [ 1034.722504] env[62558]: DEBUG nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.722836] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1034.722836] env[62558]: value = "task-1267298" [ 1034.722836] env[62558]: _type = "Task" [ 1034.722836] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.723547] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eadfe27-42a3-42fb-b476-e66b3c939576 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.737678] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267298, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.799415] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.799863] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.800226] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.822942] env[62558]: DEBUG oslo_vmware.api [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144563} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.823305] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.823529] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.823765] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.823950] env[62558]: INFO nova.compute.manager [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1034.824246] env[62558]: DEBUG oslo.service.loopingcall [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.824524] env[62558]: DEBUG nova.compute.manager [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.824627] env[62558]: DEBUG nova.network.neutron [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1034.832759] env[62558]: INFO nova.scheduler.client.report [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Deleted allocations for instance 9b87ed06-b062-4e6f-9dfa-a9b5491fe101 [ 1035.250820] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267298, 'name': ReconfigVM_Task, 'duration_secs': 0.297012} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.257222] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.257642] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.260023] env[62558]: INFO nova.compute.claims [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.265511] env[62558]: INFO nova.compute.manager [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Took 16.16 seconds to build instance. [ 1035.269382] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e7569632-751d-4bd7-b00a-5176c816551c/e7569632-751d-4bd7-b00a-5176c816551c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.269382] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b270710-d444-4cfe-9d5b-bdd11e9729b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.276611] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1035.276611] env[62558]: value = "task-1267299" [ 1035.276611] env[62558]: _type = "Task" [ 1035.276611] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.293507] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267299, 'name': Rename_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.341091] env[62558]: DEBUG oslo_concurrency.lockutils [None req-64511b16-2201-4800-8d27-8697a48c8eed tempest-DeleteServersTestJSON-1913298758 tempest-DeleteServersTestJSON-1913298758-project-member] Lock "9b87ed06-b062-4e6f-9dfa-a9b5491fe101" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.447s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.381618] env[62558]: DEBUG nova.compute.manager [req-99a02890-5370-437a-867a-6eb4339e344c req-15dc7cae-26a6-40e5-8cf0-0359cd598202 service nova] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Received event network-vif-deleted-5bb4afdb-999e-446a-ab6e-1bd7a8bc092b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.381898] env[62558]: DEBUG nova.compute.manager [req-99a02890-5370-437a-867a-6eb4339e344c req-15dc7cae-26a6-40e5-8cf0-0359cd598202 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Received event network-vif-deleted-b2ae5c23-49aa-4034-b884-b1017e034d61 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.382112] env[62558]: INFO nova.compute.manager [req-99a02890-5370-437a-867a-6eb4339e344c req-15dc7cae-26a6-40e5-8cf0-0359cd598202 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Neutron deleted interface b2ae5c23-49aa-4034-b884-b1017e034d61; detaching it from the instance and deleting it from the info cache [ 1035.382311] env[62558]: DEBUG nova.network.neutron [req-99a02890-5370-437a-867a-6eb4339e344c req-15dc7cae-26a6-40e5-8cf0-0359cd598202 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.756713] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "4dffea0b-adb7-4c6a-b2b0-436796111958" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.757008] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.757249] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.757449] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.757631] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.760904] env[62558]: INFO nova.compute.manager [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Terminating instance [ 1035.769089] env[62558]: DEBUG nova.compute.manager [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.769303] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.770174] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d240045-b13a-4eea-a666-586b44835f3a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.775112] env[62558]: DEBUG oslo_concurrency.lockutils [None req-80eab30b-8d0c-446f-a96d-d8f162d6cf57 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.681s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.778132] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.783239] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5291cefb-7c79-4a30-b6df-2a9b6d97e59c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.790151] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267299, 'name': Rename_Task, 'duration_secs': 0.142539} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.790415] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.790646] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e238d47-c8c9-439b-9d5c-17b48681a9d8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.794554] env[62558]: DEBUG nova.network.neutron [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.796655] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1035.796655] env[62558]: value = "task-1267301" [ 1035.796655] env[62558]: _type = "Task" [ 1035.796655] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.804815] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267301, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.855613] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.855844] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.856040] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleting the datastore file [datastore1] 4dffea0b-adb7-4c6a-b2b0-436796111958 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.856323] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da80da51-4139-4b5a-ab9e-eeeb565cfe54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.863528] env[62558]: DEBUG oslo_vmware.api [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for the task: (returnval){ [ 1035.863528] env[62558]: value = "task-1267302" [ 1035.863528] env[62558]: _type = "Task" [ 1035.863528] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.875827] env[62558]: DEBUG oslo_vmware.api [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.884518] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0c7c4d7-80cc-486b-9a93-1b9dea2950f9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.892472] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b938736a-2406-4396-9a90-926853817e94 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.919019] env[62558]: DEBUG nova.compute.manager [req-99a02890-5370-437a-867a-6eb4339e344c req-15dc7cae-26a6-40e5-8cf0-0359cd598202 service nova] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Detach interface failed, port_id=b2ae5c23-49aa-4034-b884-b1017e034d61, reason: Instance ea680004-b94a-47f9-b516-2afb95c30a51 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1035.930430] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1035.930706] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272614', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'name': 'volume-cee3b8af-c539-4269-8100-6e3f68136336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'serial': 'cee3b8af-c539-4269-8100-6e3f68136336'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1035.931626] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514fb465-221c-4ee3-8dc2-774cc3c04ad7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.948357] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe2a0bd-4083-4fb0-bd39-9ff7c738aafe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.972398] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-cee3b8af-c539-4269-8100-6e3f68136336/volume-cee3b8af-c539-4269-8100-6e3f68136336.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.972646] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0c1761a-d38a-49d8-a5b3-c0bdd6c0b78e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.991590] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1035.991590] env[62558]: value = "task-1267303" [ 1035.991590] env[62558]: _type = "Task" [ 1035.991590] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.999801] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267303, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.298496] env[62558]: INFO nova.compute.manager [-] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Took 1.47 seconds to deallocate network for instance. [ 1036.312614] env[62558]: DEBUG oslo_vmware.api [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267301, 'name': PowerOnVM_Task, 'duration_secs': 0.50346} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.315686] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.315924] env[62558]: INFO nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1036.316195] env[62558]: DEBUG nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.317198] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6988cbf7-391f-4726-af47-3205b5cbb8ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.373920] env[62558]: DEBUG oslo_vmware.api [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Task: {'id': task-1267302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168868} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.376518] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.376767] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.376960] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.377151] env[62558]: INFO nova.compute.manager [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1036.377390] env[62558]: DEBUG oslo.service.loopingcall [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.377777] env[62558]: DEBUG nova.compute.manager [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.377877] env[62558]: DEBUG nova.network.neutron [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.418654] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaef84b-7541-4176-80cd-bb3b85d39acd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.430866] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ba8a3b-3ec9-43e5-8baa-2285ec9284fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.466387] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9602d0-8e47-4eb1-a28d-1d741274328f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.474055] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7ec4ba-23b7-4265-a91c-0e9f7ad412e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.487196] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1036.500639] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267303, 'name': ReconfigVM_Task, 'duration_secs': 0.361375} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.500907] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-cee3b8af-c539-4269-8100-6e3f68136336/volume-cee3b8af-c539-4269-8100-6e3f68136336.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.505765] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ae78d60-a099-4866-a1b0-ad333ae2796b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.521014] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1036.521014] env[62558]: value = "task-1267304" [ 1036.521014] env[62558]: _type = "Task" [ 1036.521014] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.531569] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267304, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.810577] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.840945] env[62558]: INFO nova.compute.manager [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Took 16.81 seconds to build instance. [ 1037.013820] env[62558]: ERROR nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [req-5af47eec-73b2-484f-a456-e78e67a9de86] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5af47eec-73b2-484f-a456-e78e67a9de86"}]} [ 1037.036057] env[62558]: DEBUG nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1037.042356] env[62558]: DEBUG oslo_vmware.api [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267304, 'name': ReconfigVM_Task, 'duration_secs': 0.190415} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.042860] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272614', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'name': 'volume-cee3b8af-c539-4269-8100-6e3f68136336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'serial': 'cee3b8af-c539-4269-8100-6e3f68136336'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1037.056397] env[62558]: DEBUG nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1037.056397] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.070284] env[62558]: DEBUG nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1037.105025] env[62558]: DEBUG nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1037.105025] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 130 to 131 during operation: update_traits {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1037.263374] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5d9c15-0267-44b1-875c-d3a771b2af9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.272322] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c29ee6-4e50-42b4-b4ab-3c8b8629294c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.317803] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b6eeca-a595-4a59-95c4-bd6f1a2c7c65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.328057] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ea88b0-ffe5-483f-83e6-91752c0bb813 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.345764] env[62558]: DEBUG nova.network.neutron [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.348063] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d8a17807-3333-48dc-8df3-885d04d01702 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.329s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.348545] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.410089] env[62558]: DEBUG nova.compute.manager [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.410393] env[62558]: DEBUG nova.compute.manager [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.410740] env[62558]: DEBUG oslo_concurrency.lockutils [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.410964] env[62558]: DEBUG oslo_concurrency.lockutils [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.411222] env[62558]: DEBUG nova.network.neutron [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.554917] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.555545] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.852044] env[62558]: INFO nova.compute.manager [-] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Took 1.47 seconds to deallocate network for instance. [ 1037.883758] env[62558]: DEBUG nova.scheduler.client.report [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1037.884055] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 131 to 132 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1037.884586] env[62558]: DEBUG nova.compute.provider_tree [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1037.931159] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "e7569632-751d-4bd7-b00a-5176c816551c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.931405] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.931609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "e7569632-751d-4bd7-b00a-5176c816551c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.932215] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.932403] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.934256] env[62558]: INFO nova.compute.manager [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Terminating instance [ 1037.935939] env[62558]: DEBUG nova.compute.manager [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.936152] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.936994] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854c96b0-5292-4fed-9f1e-74bfe5686467 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.945618] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.945887] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c597e89-6362-4244-a8ff-24cbdf4a1a0d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.951894] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1037.951894] env[62558]: value = "task-1267305" [ 1037.951894] env[62558]: _type = "Task" [ 1037.951894] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.961167] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.060584] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.101862] env[62558]: DEBUG nova.objects.instance [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.185190] env[62558]: DEBUG nova.network.neutron [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.185190] env[62558]: DEBUG nova.network.neutron [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.360723] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.389773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.132s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.390344] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.392851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.584s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.393115] env[62558]: DEBUG nova.objects.instance [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'resources' on Instance uuid ea680004-b94a-47f9-b516-2afb95c30a51 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.462460] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267305, 'name': PowerOffVM_Task, 'duration_secs': 0.273912} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.462816] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.462969] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.463250] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-057626c3-e948-4f89-a1ed-ca7f0eb37480 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.525946] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.526099] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.526294] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Deleting the datastore file [datastore1] e7569632-751d-4bd7-b00a-5176c816551c {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.526569] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a93f659-b940-482a-aed7-16064a8254e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.532390] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for the task: (returnval){ [ 1038.532390] env[62558]: value = "task-1267307" [ 1038.532390] env[62558]: _type = "Task" [ 1038.532390] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.539712] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.586792] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.607661] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9f404d88-8934-4b4d-bbe3-5198d18dc728 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.278s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.687698] env[62558]: DEBUG oslo_concurrency.lockutils [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.687999] env[62558]: DEBUG nova.compute.manager [req-59b43599-7cb1-4d15-8ccb-f86087fdf8f2 req-562dc4d8-82aa-4c61-a84f-70092f870cc0 service nova] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Received event network-vif-deleted-68f78b29-4c73-4a2c-b92f-1d34245b9c68 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.895854] env[62558]: DEBUG nova.compute.utils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.901066] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1038.901066] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.955139] env[62558]: DEBUG nova.policy [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e6e2f1bae4d4bb68070770dbbbe5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78b69738b90142a4943069006e349445', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.031266] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16158ab-0266-4f44-a5de-c29ea41c006c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.053697] env[62558]: DEBUG oslo_vmware.api [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Task: {'id': task-1267307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150758} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.055155] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf452b2-a15b-47da-805b-0b5370a00e25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.064558] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.065157] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.065718] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.065930] env[62558]: INFO nova.compute.manager [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1039.066201] env[62558]: DEBUG oslo.service.loopingcall [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.066418] env[62558]: DEBUG nova.compute.manager [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.066581] env[62558]: DEBUG nova.network.neutron [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.106227] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f17011-460e-44e2-ad56-7ad52bea0d58 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.115900] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72209a3-3326-4f51-a19a-801368378861 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.129239] env[62558]: DEBUG nova.compute.provider_tree [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.317591] env[62558]: DEBUG nova.compute.manager [req-4ccfc7d3-8fac-4658-8bb0-5d5928422a45 req-33c7a53e-1503-45e8-bca0-b99096ce8e7a service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Received event network-vif-deleted-2e27a73e-589c-4e74-bfc4-23d8a521c175 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.317796] env[62558]: INFO nova.compute.manager [req-4ccfc7d3-8fac-4658-8bb0-5d5928422a45 req-33c7a53e-1503-45e8-bca0-b99096ce8e7a service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Neutron deleted interface 2e27a73e-589c-4e74-bfc4-23d8a521c175; detaching it from the instance and deleting it from the info cache [ 1039.317971] env[62558]: DEBUG nova.network.neutron [req-4ccfc7d3-8fac-4658-8bb0-5d5928422a45 req-33c7a53e-1503-45e8-bca0-b99096ce8e7a service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.392974] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Successfully created port: fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.403516] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.501184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.501712] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.002s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.635132] env[62558]: DEBUG nova.scheduler.client.report [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.793106] env[62558]: DEBUG nova.network.neutron [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.820824] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ace0dd96-3fe3-426c-9f49-776df73f08de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.834026] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0370d734-e8b0-4a6b-b742-3d0ad0307e34 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.858056] env[62558]: DEBUG nova.compute.manager [req-4ccfc7d3-8fac-4658-8bb0-5d5928422a45 req-33c7a53e-1503-45e8-bca0-b99096ce8e7a service nova] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Detach interface failed, port_id=2e27a73e-589c-4e74-bfc4-23d8a521c175, reason: Instance e7569632-751d-4bd7-b00a-5176c816551c could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1040.009578] env[62558]: DEBUG nova.compute.utils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.140866] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.147169] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.786s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.147169] env[62558]: DEBUG nova.objects.instance [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lazy-loading 'resources' on Instance uuid 4dffea0b-adb7-4c6a-b2b0-436796111958 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.169199] env[62558]: INFO nova.scheduler.client.report [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted allocations for instance ea680004-b94a-47f9-b516-2afb95c30a51 [ 1040.297210] env[62558]: INFO nova.compute.manager [-] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Took 1.23 seconds to deallocate network for instance. [ 1040.413034] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.440224] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.440224] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.440224] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.440224] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.440515] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.440846] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.442017] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.442327] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.442615] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.442894] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.443375] env[62558]: DEBUG nova.virt.hardware [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.444484] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a83e10-9932-4569-a9f6-6a3c710ffc18 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.453666] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1cbcc3-99f2-4e1f-a1a4-9ef37d2fc84e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.511809] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.675395] env[62558]: DEBUG oslo_concurrency.lockutils [None req-88fd0442-70b9-460d-9549-76667d180c92 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "ea680004-b94a-47f9-b516-2afb95c30a51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.073s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.752431] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6df8b5-92d2-4c85-a607-c1e06f82caa3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.759631] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3196360-b83e-4937-8a5e-e50ea9a1de92 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.790551] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc48a966-6dde-4bbd-a6b0-0fe86915064a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.797904] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15133c24-5c89-4f34-adfc-0425c4640a16 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.802530] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.812252] env[62558]: DEBUG nova.compute.provider_tree [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.078410] env[62558]: DEBUG nova.compute.manager [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Received event network-vif-plugged-fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.078642] env[62558]: DEBUG oslo_concurrency.lockutils [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] Acquiring lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.078859] env[62558]: DEBUG oslo_concurrency.lockutils [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.079716] env[62558]: DEBUG oslo_concurrency.lockutils [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.082332] env[62558]: DEBUG nova.compute.manager [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] No waiting events found dispatching network-vif-plugged-fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.082550] env[62558]: WARNING nova.compute.manager [req-e86e6780-f0af-44b3-a629-1956796954a1 req-e65e6953-1690-4137-86ad-3466a07588b7 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Received unexpected event network-vif-plugged-fbef2b6a-ad6d-410d-9495-f3c25b3d937a for instance with vm_state building and task_state spawning. [ 1041.253551] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Successfully updated port: fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.316013] env[62558]: DEBUG nova.scheduler.client.report [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.614423] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.614773] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.614996] env[62558]: INFO nova.compute.manager [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attaching volume d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0 to /dev/sdc [ 1041.652023] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e943b5-e4ad-4606-b394-d304caaf241a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.660244] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e390ee44-ce76-4356-9f82-4ce194389229 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.674802] env[62558]: DEBUG nova.virt.block_device [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating existing volume attachment record: b7864215-a41e-41dd-b602-06ad0c11f4a9 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1041.757829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.757829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.757829] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.825580] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.827584] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.241s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.829453] env[62558]: INFO nova.compute.claims [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.859647] env[62558]: INFO nova.scheduler.client.report [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Deleted allocations for instance 4dffea0b-adb7-4c6a-b2b0-436796111958 [ 1041.997316] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.997691] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.291147] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.368413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-78edae7f-c3be-47d1-a392-218b8a6e6861 tempest-ServersTestJSON-2143194746 tempest-ServersTestJSON-2143194746-project-member] Lock "4dffea0b-adb7-4c6a-b2b0-436796111958" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.611s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.446189] env[62558]: DEBUG nova.network.neutron [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updating instance_info_cache with network_info: [{"id": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "address": "fa:16:3e:05:55:7a", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbef2b6a-ad", "ovs_interfaceid": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.455194] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.455446] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.501114] env[62558]: INFO nova.compute.manager [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Detaching volume 4d0add79-8ef3-4cf7-be38-7fd542cda129 [ 1042.535400] env[62558]: INFO nova.virt.block_device [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Attempting to driver detach volume 4d0add79-8ef3-4cf7-be38-7fd542cda129 from mountpoint /dev/sdb [ 1042.535695] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1042.535949] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272598', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'name': 'volume-4d0add79-8ef3-4cf7-be38-7fd542cda129', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3f42034-3d2b-472a-89c0-5445cb6fb567', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'serial': '4d0add79-8ef3-4cf7-be38-7fd542cda129'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1042.536912] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1022c0-d9a4-4a0c-b641-2564b7a8951b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.559235] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e35706d-c5f9-40fd-81e9-840cd210fd6a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.566301] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8420cbd-3552-4519-96f6-c7c5db6ee70c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.588307] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533f7f36-7cd3-4eb3-8b5d-8b2527d6dc98 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.603211] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] The volume has not been displaced from its original location: [datastore1] volume-4d0add79-8ef3-4cf7-be38-7fd542cda129/volume-4d0add79-8ef3-4cf7-be38-7fd542cda129.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1042.608534] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1042.608857] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b39180b-8494-41e0-8efb-1f889a8fd2ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.626946] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1042.626946] env[62558]: value = "task-1267309" [ 1042.626946] env[62558]: _type = "Task" [ 1042.626946] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.634599] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.947330] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8670efe-aede-4f1a-8771-ddf061af8071 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.950419] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.950808] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Instance network_info: |[{"id": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "address": "fa:16:3e:05:55:7a", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbef2b6a-ad", "ovs_interfaceid": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.951330] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:55:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbef2b6a-ad6d-410d-9495-f3c25b3d937a', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.958964] env[62558]: DEBUG oslo.service.loopingcall [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.959719] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1042.962292] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.965311] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-789e0b93-86dd-42d3-9e94-f6b07a01067c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.980681] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9695bf0d-bb51-4565-b6a8-cc0527aa4dfd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.012465] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.012465] env[62558]: value = "task-1267310" [ 1043.012465] env[62558]: _type = "Task" [ 1043.012465] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.013540] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df006ea2-cefb-4cef-81a4-53a22aa339f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.027687] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e713a11-1215-4053-8fb2-0ad5645cd6a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.031543] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267310, 'name': CreateVM_Task} progress is 15%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.045020] env[62558]: DEBUG nova.compute.provider_tree [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.113080] env[62558]: DEBUG nova.compute.manager [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Received event network-changed-fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.113388] env[62558]: DEBUG nova.compute.manager [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Refreshing instance network info cache due to event network-changed-fbef2b6a-ad6d-410d-9495-f3c25b3d937a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.113631] env[62558]: DEBUG oslo_concurrency.lockutils [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] Acquiring lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.113787] env[62558]: DEBUG oslo_concurrency.lockutils [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] Acquired lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.113931] env[62558]: DEBUG nova.network.neutron [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Refreshing network info cache for port fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1043.137788] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267309, 'name': ReconfigVM_Task, 'duration_secs': 0.249521} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.138071] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1043.142731] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8add2c09-aa5b-42ac-8b74-355a45e8b192 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.161040] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1043.161040] env[62558]: value = "task-1267311" [ 1043.161040] env[62558]: _type = "Task" [ 1043.161040] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.169501] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267311, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.478237] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.522922] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267310, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.547083] env[62558]: DEBUG nova.scheduler.client.report [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.670336] env[62558]: DEBUG oslo_vmware.api [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267311, 'name': ReconfigVM_Task, 'duration_secs': 0.144995} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.670591] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272598', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'name': 'volume-4d0add79-8ef3-4cf7-be38-7fd542cda129', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a3f42034-3d2b-472a-89c0-5445cb6fb567', 'attached_at': '', 'detached_at': '', 'volume_id': '4d0add79-8ef3-4cf7-be38-7fd542cda129', 'serial': '4d0add79-8ef3-4cf7-be38-7fd542cda129'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1043.941338] env[62558]: DEBUG nova.network.neutron [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updated VIF entry in instance network info cache for port fbef2b6a-ad6d-410d-9495-f3c25b3d937a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.942052] env[62558]: DEBUG nova.network.neutron [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updating instance_info_cache with network_info: [{"id": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "address": "fa:16:3e:05:55:7a", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbef2b6a-ad", "ovs_interfaceid": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.026490] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267310, 'name': CreateVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.052221] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.052745] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.055629] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.253s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.055850] env[62558]: DEBUG nova.objects.instance [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lazy-loading 'resources' on Instance uuid e7569632-751d-4bd7-b00a-5176c816551c {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.219575] env[62558]: DEBUG nova.objects.instance [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid a3f42034-3d2b-472a-89c0-5445cb6fb567 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.448507] env[62558]: DEBUG oslo_concurrency.lockutils [req-d0269ad3-b96e-49e2-81eb-efb504e18b7d req-9d4070c5-140b-4114-9119-020a534f18d1 service nova] Releasing lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.529260] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267310, 'name': CreateVM_Task, 'duration_secs': 1.38371} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.529464] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.530243] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.530412] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.530751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.531048] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0f330b8-efc0-4687-a9e3-50e06c516c22 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.535473] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1044.535473] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eea0bf-e543-9c06-31ec-748396b13340" [ 1044.535473] env[62558]: _type = "Task" [ 1044.535473] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.543280] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eea0bf-e543-9c06-31ec-748396b13340, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.558540] env[62558]: DEBUG nova.compute.utils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.562713] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.562766] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.632281] env[62558]: DEBUG nova.policy [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.703692] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63765410-f69d-4d9e-8d16-e227af20b1a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.714051] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922ef934-dffb-4def-9196-03d6c3fdaadc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.749382] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b4676e-17e6-4121-812f-e8bd5d9fe240 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.758230] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c4e1af-fe3e-42a3-8623-4d0039117b92 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.775526] env[62558]: DEBUG nova.compute.provider_tree [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1044.960081] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Successfully created port: 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.048534] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52eea0bf-e543-9c06-31ec-748396b13340, 'name': SearchDatastore_Task, 'duration_secs': 0.008866} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.048786] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.049077] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.049445] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.049994] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.049994] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.050468] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2644fb8e-ee64-4934-aadf-2bb029a1f43b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.061303] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.061303] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1045.061303] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a0ecd53-979c-4be4-839c-fe5b3c6b67fe {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.064909] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1045.072379] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1045.072379] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52713493-63c2-3b57-9711-ed0b10c993bd" [ 1045.072379] env[62558]: _type = "Task" [ 1045.072379] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.081172] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52713493-63c2-3b57-9711-ed0b10c993bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.253782] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9eddf27f-9b3c-435d-97e4-ba9f0ad8ab8e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.256s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.294403] env[62558]: ERROR nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] [req-6460cfe8-b1e0-4057-8f20-56ed58c6e398] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6460cfe8-b1e0-4057-8f20-56ed58c6e398"}]} [ 1045.309819] env[62558]: DEBUG nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1045.322131] env[62558]: DEBUG nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1045.322339] env[62558]: DEBUG nova.compute.provider_tree [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.331994] env[62558]: DEBUG nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1045.347595] env[62558]: DEBUG nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1045.437135] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e9cc23-950a-4a25-9ea9-f63cd0ba18d7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.446568] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd5afcb-cd5d-4ef0-ae26-cda7b95b11d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.475049] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c28f95-2d88-4842-b66c-abd5f4c25ea2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.481665] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca83fbca-3750-4170-99f0-a34e3857795c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.494355] env[62558]: DEBUG nova.compute.provider_tree [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1045.582275] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52713493-63c2-3b57-9711-ed0b10c993bd, 'name': SearchDatastore_Task, 'duration_secs': 0.008486} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.583119] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da7549fa-b208-498c-85d8-e65b576f7432 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.587909] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1045.587909] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb8d15-d658-e2a2-2b2d-7c200cb5a7cb" [ 1045.587909] env[62558]: _type = "Task" [ 1045.587909] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.594981] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb8d15-d658-e2a2-2b2d-7c200cb5a7cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.028953] env[62558]: DEBUG nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1046.029390] env[62558]: DEBUG nova.compute.provider_tree [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 133 to 134 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1046.029453] env[62558]: DEBUG nova.compute.provider_tree [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1046.075045] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.098878] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb8d15-d658-e2a2-2b2d-7c200cb5a7cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009255} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.099176] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.099503] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 5082012b-e8ed-41d6-a036-b64cbb1e363c/5082012b-e8ed-41d6-a036-b64cbb1e363c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1046.099709] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7680066e-8347-4ea8-b79f-5341592706b7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.104355] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.104585] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.104743] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.104926] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.105085] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.105257] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.105494] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.105661] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.105829] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.105993] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.106187] env[62558]: DEBUG nova.virt.hardware [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.107816] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35c0eef-a11b-4af3-9c00-db6198fc6fac {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.110699] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1046.110699] env[62558]: value = "task-1267313" [ 1046.110699] env[62558]: _type = "Task" [ 1046.110699] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.117611] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f052ed-4824-45eb-be32-3d3d90c1c43b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.124395] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.223120] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1046.223658] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272616', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'name': 'volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'serial': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1046.224338] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4676e228-94b1-469b-be89-ecb452346829 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.242306] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a5b9ab-c6ce-4125-acf2-67a6c08fdf0b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.263424] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.263718] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.263935] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.264159] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.264349] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.273362] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0/volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.274195] env[62558]: INFO nova.compute.manager [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Terminating instance [ 1046.275623] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e9b3515-598f-4e35-b8c9-8dc92b293221 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.290297] env[62558]: DEBUG nova.compute.manager [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1046.290526] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1046.291440] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a06d0d-5c74-4330-a3e2-40db634790bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.299237] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.299989] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65b82007-b142-43f0-8032-3050fe22294c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.302386] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1046.302386] env[62558]: value = "task-1267314" [ 1046.302386] env[62558]: _type = "Task" [ 1046.302386] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.308404] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1046.308404] env[62558]: value = "task-1267315" [ 1046.308404] env[62558]: _type = "Task" [ 1046.308404] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.316294] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267314, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.322968] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.398460] env[62558]: DEBUG nova.compute.manager [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-vif-plugged-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.398784] env[62558]: DEBUG oslo_concurrency.lockutils [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.399449] env[62558]: DEBUG oslo_concurrency.lockutils [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.399449] env[62558]: DEBUG oslo_concurrency.lockutils [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.399449] env[62558]: DEBUG nova.compute.manager [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] No waiting events found dispatching network-vif-plugged-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.399574] env[62558]: WARNING nova.compute.manager [req-40f7d21e-5be8-48b7-a3cb-610d612ec43a req-578a6302-0211-4ef1-a5ff-1fb65aac5677 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received unexpected event network-vif-plugged-66f313e4-07aa-4e57-b472-a4408da24e26 for instance with vm_state building and task_state spawning. [ 1046.517857] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Successfully updated port: 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.534182] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.478s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.536989] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.059s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.541963] env[62558]: INFO nova.compute.claims [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.561202] env[62558]: INFO nova.scheduler.client.report [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Deleted allocations for instance e7569632-751d-4bd7-b00a-5176c816551c [ 1046.621180] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482653} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.621532] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 5082012b-e8ed-41d6-a036-b64cbb1e363c/5082012b-e8ed-41d6-a036-b64cbb1e363c.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.621779] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.622048] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb722ed7-4dc2-4302-b54b-83ee1174552e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.627585] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1046.627585] env[62558]: value = "task-1267316" [ 1046.627585] env[62558]: _type = "Task" [ 1046.627585] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.635243] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.815197] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267314, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.820144] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267315, 'name': PowerOffVM_Task, 'duration_secs': 0.208807} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.820410] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.820584] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1046.820831] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0df7e1a9-ced7-4ca3-b570-d22c63116f93 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.902103] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.902365] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.902555] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleting the datastore file [datastore2] a3f42034-3d2b-472a-89c0-5445cb6fb567 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.902831] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bff5d71-d841-4444-a225-6be69db3e8ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.909823] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1046.909823] env[62558]: value = "task-1267318" [ 1046.909823] env[62558]: _type = "Task" [ 1046.909823] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.917980] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.021140] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.021337] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.021454] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1047.068726] env[62558]: DEBUG oslo_concurrency.lockutils [None req-d2d53926-950d-4baf-98a7-f80b2a3b37a7 tempest-ServersNegativeTestMultiTenantJSON-21731487 tempest-ServersNegativeTestMultiTenantJSON-21731487-project-member] Lock "e7569632-751d-4bd7-b00a-5176c816551c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.137s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.136905] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065751} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.137192] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.137956] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f844a5-f22f-4647-a1dd-a6aac51be42a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.159629] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 5082012b-e8ed-41d6-a036-b64cbb1e363c/5082012b-e8ed-41d6-a036-b64cbb1e363c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.159928] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdb779e2-589d-42b2-bd24-14313b80f19a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.179065] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1047.179065] env[62558]: value = "task-1267319" [ 1047.179065] env[62558]: _type = "Task" [ 1047.179065] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.187246] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.314062] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267314, 'name': ReconfigVM_Task, 'duration_secs': 0.590435} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.314290] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0/volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.319283] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c01f2c44-979c-43ef-bfb6-bfc6b56a5864 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.336561] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1047.336561] env[62558]: value = "task-1267320" [ 1047.336561] env[62558]: _type = "Task" [ 1047.336561] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.345620] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.420585] env[62558]: DEBUG oslo_vmware.api [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150698} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.420915] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.421125] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1047.421309] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1047.421488] env[62558]: INFO nova.compute.manager [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1047.421743] env[62558]: DEBUG oslo.service.loopingcall [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.421995] env[62558]: DEBUG nova.compute.manager [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.422107] env[62558]: DEBUG nova.network.neutron [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1047.555110] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.674376] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a6a307-ee88-4175-ae6a-b05cbb30eeaf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.684844] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3052441-ba6b-4e94-9684-716d1825445a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.693479] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267319, 'name': ReconfigVM_Task, 'duration_secs': 0.345387} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.717416] env[62558]: DEBUG nova.network.neutron [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.718712] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 5082012b-e8ed-41d6-a036-b64cbb1e363c/5082012b-e8ed-41d6-a036-b64cbb1e363c.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.723830] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7019a46-9902-4e03-952d-30eecd78766b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.726011] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d274efa1-9abb-43f7-9edd-876de8b63272 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.738524] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97219c6-0da4-4313-963c-ae2c26b118db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.743356] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1047.743356] env[62558]: value = "task-1267321" [ 1047.743356] env[62558]: _type = "Task" [ 1047.743356] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.755028] env[62558]: DEBUG nova.compute.provider_tree [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.763537] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267321, 'name': Rename_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.847567] env[62558]: DEBUG oslo_vmware.api [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267320, 'name': ReconfigVM_Task, 'duration_secs': 0.147943} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.847959] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272616', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'name': 'volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'serial': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.224746] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.225153] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Instance network_info: |[{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.225681] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:5d:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '715e3f37-7401-48fb-a0ee-59d340b40de1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66f313e4-07aa-4e57-b472-a4408da24e26', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.233282] env[62558]: DEBUG oslo.service.loopingcall [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.233588] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1048.233744] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b51c21b5-465f-4b62-ad59-d130a205bb7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.258448] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267321, 'name': Rename_Task, 'duration_secs': 0.13765} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.259795] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1048.260789] env[62558]: DEBUG nova.scheduler.client.report [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.263527] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.263527] env[62558]: value = "task-1267322" [ 1048.263527] env[62558]: _type = "Task" [ 1048.263527] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.263720] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ac6875a-4f41-480d-a65e-ce37e17306cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.276344] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267322, 'name': CreateVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.277981] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1048.277981] env[62558]: value = "task-1267323" [ 1048.277981] env[62558]: _type = "Task" [ 1048.277981] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.286295] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.429644] env[62558]: DEBUG nova.compute.manager [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.430059] env[62558]: DEBUG nova.compute.manager [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing instance network info cache due to event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.430162] env[62558]: DEBUG oslo_concurrency.lockutils [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.430499] env[62558]: DEBUG oslo_concurrency.lockutils [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.430883] env[62558]: DEBUG nova.network.neutron [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.633795] env[62558]: DEBUG nova.network.neutron [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.769808] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.770560] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.791375] env[62558]: DEBUG oslo_vmware.api [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267323, 'name': PowerOnVM_Task, 'duration_secs': 0.501674} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.794845] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.795384] env[62558]: INFO nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Took 8.38 seconds to spawn the instance on the hypervisor. [ 1048.795384] env[62558]: DEBUG nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.795578] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267322, 'name': CreateVM_Task, 'duration_secs': 0.398189} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.796352] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174196c5-faa9-41a2-ad16-b6e6d6147c97 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.799097] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.799760] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.800228] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.800336] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.800922] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23c1fa28-e972-46ce-a609-92f5cafd76b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.811023] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1048.811023] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5200153e-c21d-ad45-80ae-df9bc3e72792" [ 1048.811023] env[62558]: _type = "Task" [ 1048.811023] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.818972] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5200153e-c21d-ad45-80ae-df9bc3e72792, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.904183] env[62558]: DEBUG nova.objects.instance [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.136746] env[62558]: INFO nova.compute.manager [-] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Took 1.71 seconds to deallocate network for instance. [ 1049.283551] env[62558]: DEBUG nova.compute.utils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1049.284975] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1049.285158] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1049.318684] env[62558]: INFO nova.compute.manager [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Took 14.09 seconds to build instance. [ 1049.328460] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5200153e-c21d-ad45-80ae-df9bc3e72792, 'name': SearchDatastore_Task, 'duration_secs': 0.016838} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.329216] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.329719] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.330009] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.330253] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.331336] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.331511] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e237674-81a7-48d3-a7dc-f536fbc2b234 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.345829] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.346040] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1049.350140] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8508adb5-2b8a-4ff8-8fc2-cde259dffdce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.357638] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1049.357638] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aa25e8-2787-8004-ed33-bb3387a9cfc8" [ 1049.357638] env[62558]: _type = "Task" [ 1049.357638] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.367888] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aa25e8-2787-8004-ed33-bb3387a9cfc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.369604] env[62558]: DEBUG nova.policy [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '985bef0c1e7e429e986753953b597926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '688bb982a71a4730aa1652ea837e1b73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1049.408937] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e1ba8154-cfde-41e4-a4b4-5fb140f707a5 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.794s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.480646] env[62558]: DEBUG nova.network.neutron [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updated VIF entry in instance network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1049.481035] env[62558]: DEBUG nova.network.neutron [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.629640] env[62558]: DEBUG nova.compute.manager [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Received event network-changed-fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.629862] env[62558]: DEBUG nova.compute.manager [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Refreshing instance network info cache due to event network-changed-fbef2b6a-ad6d-410d-9495-f3c25b3d937a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1049.630754] env[62558]: DEBUG oslo_concurrency.lockutils [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] Acquiring lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.631109] env[62558]: DEBUG oslo_concurrency.lockutils [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] Acquired lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.631775] env[62558]: DEBUG nova.network.neutron [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Refreshing network info cache for port fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1049.643308] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.643674] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.644014] env[62558]: DEBUG nova.objects.instance [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'resources' on Instance uuid a3f42034-3d2b-472a-89c0-5445cb6fb567 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.790664] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1049.806770] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Successfully created port: 44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1049.820792] env[62558]: DEBUG oslo_concurrency.lockutils [None req-1b84d930-e805-4af5-bc34-78689890ecbe tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.611s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.867707] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aa25e8-2787-8004-ed33-bb3387a9cfc8, 'name': SearchDatastore_Task, 'duration_secs': 0.042331} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.868483] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-812d5622-d28b-481c-baa8-05a4bdd9fce8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.873715] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1049.873715] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e21e6-7a88-08d0-ed6a-96922a25df48" [ 1049.873715] env[62558]: _type = "Task" [ 1049.873715] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.883086] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e21e6-7a88-08d0-ed6a-96922a25df48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.986501] env[62558]: DEBUG oslo_concurrency.lockutils [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.986501] env[62558]: DEBUG nova.compute.manager [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Received event network-vif-deleted-8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.986501] env[62558]: INFO nova.compute.manager [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Neutron deleted interface 8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a; detaching it from the instance and deleting it from the info cache [ 1049.986501] env[62558]: DEBUG nova.network.neutron [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.268112] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146b99fd-0290-4041-b2e2-b20bf7c4976f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.277414] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6824bcf-ce9f-4e25-81c6-cb17bb694a20 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.316992] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c0fcc3-4f84-4366-b0f0-ff2469b33940 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.320851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.322022] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.332015] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1ff0aa-a40a-4e85-a324-6e72e33098c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.345555] env[62558]: DEBUG nova.compute.provider_tree [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.384118] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]525e21e6-7a88-08d0-ed6a-96922a25df48, 'name': SearchDatastore_Task, 'duration_secs': 0.010721} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.387622] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.388055] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ac1ec03d-e04b-42eb-a1f3-c14887241bfa/ac1ec03d-e04b-42eb-a1f3-c14887241bfa.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1050.388473] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4907793-785f-47fe-9959-a6847007519b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.396120] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1050.396120] env[62558]: value = "task-1267324" [ 1050.396120] env[62558]: _type = "Task" [ 1050.396120] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.403931] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267324, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.464724] env[62558]: DEBUG nova.network.neutron [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updated VIF entry in instance network info cache for port fbef2b6a-ad6d-410d-9495-f3c25b3d937a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1050.464724] env[62558]: DEBUG nova.network.neutron [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updating instance_info_cache with network_info: [{"id": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "address": "fa:16:3e:05:55:7a", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbef2b6a-ad", "ovs_interfaceid": "fbef2b6a-ad6d-410d-9495-f3c25b3d937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.492020] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7e239e7-972b-491a-969c-28e20b648e85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.498637] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0610921d-3c98-4f0b-bfe7-cd46e00e3d7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.529869] env[62558]: DEBUG nova.compute.manager [req-11edc3f0-49a9-470f-8bf7-9cef41b398ca req-5e582bb8-ca65-46e6-985d-a80525ea7831 service nova] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Detach interface failed, port_id=8ff73e0e-0c2c-4121-90c1-6dc2d76e9e4a, reason: Instance a3f42034-3d2b-472a-89c0-5445cb6fb567 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1050.819562] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1050.824269] env[62558]: INFO nova.compute.manager [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Detaching volume cee3b8af-c539-4269-8100-6e3f68136336 [ 1050.851721] env[62558]: DEBUG nova.scheduler.client.report [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.858867] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.859306] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.859770] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.859770] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.859849] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.860013] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.861913] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.861913] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.861913] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.861913] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.861913] env[62558]: DEBUG nova.virt.hardware [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.862258] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6460848d-bebf-4d5f-ab79-91cf872f30ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.866973] env[62558]: INFO nova.virt.block_device [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attempting to driver detach volume cee3b8af-c539-4269-8100-6e3f68136336 from mountpoint /dev/sdb [ 1050.867212] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1050.867578] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272614', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'name': 'volume-cee3b8af-c539-4269-8100-6e3f68136336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'serial': 'cee3b8af-c539-4269-8100-6e3f68136336'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1050.868208] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adf3fb3-269e-4fda-a1ca-9f6540031868 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.874791] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db239ac-e1af-40fd-976c-0c034ad2a73a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.897821] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f650580-2712-4949-b14e-d792ef30084d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.916555] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267324, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.918775] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1393976d-c787-4190-bb3d-db817a3ee816 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.942374] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84613007-4a0a-410c-93cb-c6e00c476380 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.959465] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] The volume has not been displaced from its original location: [datastore1] volume-cee3b8af-c539-4269-8100-6e3f68136336/volume-cee3b8af-c539-4269-8100-6e3f68136336.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1050.964652] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1050.965009] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-757f25e6-e75b-4cd8-a438-00c281cb270c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.978225] env[62558]: DEBUG oslo_concurrency.lockutils [req-f2e0ca8f-76a0-44f2-8301-b68d7c03c70a req-3c89f9fe-52ae-4eae-84be-5eb06c6486cd service nova] Releasing lock "refresh_cache-5082012b-e8ed-41d6-a036-b64cbb1e363c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.984580] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1050.984580] env[62558]: value = "task-1267325" [ 1050.984580] env[62558]: _type = "Task" [ 1050.984580] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.993901] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.366398] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.402931] env[62558]: INFO nova.scheduler.client.report [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted allocations for instance a3f42034-3d2b-472a-89c0-5445cb6fb567 [ 1051.410165] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267324, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.495806] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.770363] env[62558]: DEBUG nova.compute.manager [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Received event network-vif-plugged-44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.770590] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.770794] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.770962] env[62558]: DEBUG oslo_concurrency.lockutils [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.771226] env[62558]: DEBUG nova.compute.manager [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] No waiting events found dispatching network-vif-plugged-44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1051.771445] env[62558]: WARNING nova.compute.manager [req-c47b7b39-5ba5-48b9-8983-185883f83ee9 req-d6126211-a779-4a3e-9d9c-1197a301db8b service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Received unexpected event network-vif-plugged-44bbeb25-dd9d-4cc1-914a-a3edad861b31 for instance with vm_state building and task_state spawning. [ 1051.857453] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Successfully updated port: 44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1051.908262] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267324, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.464474} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.908629] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] ac1ec03d-e04b-42eb-a1f3-c14887241bfa/ac1ec03d-e04b-42eb-a1f3-c14887241bfa.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1051.908831] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1051.909125] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27ed5f45-f8ae-414d-9e04-9e6afb0205a5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.913527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-378be49c-9997-48e0-8bc0-45f01d98be73 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "a3f42034-3d2b-472a-89c0-5445cb6fb567" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.650s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.917303] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1051.917303] env[62558]: value = "task-1267327" [ 1051.917303] env[62558]: _type = "Task" [ 1051.917303] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.925066] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.995784] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.360334] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.360495] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.360638] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.426478] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.496526] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.888423] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1052.929134] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.738415} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.930109] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1052.931667] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f98a98-7559-44f8-b3f9-8af35d215989 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.952542] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] ac1ec03d-e04b-42eb-a1f3-c14887241bfa/ac1ec03d-e04b-42eb-a1f3-c14887241bfa.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.955267] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51f346cb-b0b4-4882-9e91-4e43254b9a14 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.976993] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1052.976993] env[62558]: value = "task-1267328" [ 1052.976993] env[62558]: _type = "Task" [ 1052.976993] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.984401] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267328, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.995412] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267325, 'name': ReconfigVM_Task, 'duration_secs': 1.677699} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.995763] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1053.000485] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b124673e-a985-4f46-98a4-40c4f66d20a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.017141] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1053.017141] env[62558]: value = "task-1267329" [ 1053.017141] env[62558]: _type = "Task" [ 1053.017141] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.025057] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267329, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.049793] env[62558]: DEBUG nova.network.neutron [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.487362] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267328, 'name': ReconfigVM_Task, 'duration_secs': 0.249542} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.487737] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfigured VM instance instance-0000006b to attach disk [datastore2] ac1ec03d-e04b-42eb-a1f3-c14887241bfa/ac1ec03d-e04b-42eb-a1f3-c14887241bfa.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.488329] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fddff096-b745-4e9c-9c04-86f47eec8653 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.495682] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1053.495682] env[62558]: value = "task-1267330" [ 1053.495682] env[62558]: _type = "Task" [ 1053.495682] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.505957] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267330, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.527488] env[62558]: DEBUG oslo_vmware.api [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267329, 'name': ReconfigVM_Task, 'duration_secs': 0.127904} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.527827] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272614', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'name': 'volume-cee3b8af-c539-4269-8100-6e3f68136336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'cee3b8af-c539-4269-8100-6e3f68136336', 'serial': 'cee3b8af-c539-4269-8100-6e3f68136336'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1053.552931] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.553403] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Instance network_info: |[{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1053.553949] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:0e:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44bbeb25-dd9d-4cc1-914a-a3edad861b31', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.561450] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating folder: Project (688bb982a71a4730aa1652ea837e1b73). Parent ref: group-v272451. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1053.561739] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c95d9f20-8b35-4b24-8012-c94b44f3fc51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.574742] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created folder: Project (688bb982a71a4730aa1652ea837e1b73) in parent group-v272451. [ 1053.575132] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating folder: Instances. Parent ref: group-v272620. {{(pid=62558) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1053.575270] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2569b29f-57b8-4b4d-a28f-e04521218922 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.586596] env[62558]: INFO nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created folder: Instances in parent group-v272620. [ 1053.586859] env[62558]: DEBUG oslo.service.loopingcall [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.587071] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.587296] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6416767-c728-4117-909e-8497ea1d5878 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.605577] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.605577] env[62558]: value = "task-1267333" [ 1053.605577] env[62558]: _type = "Task" [ 1053.605577] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.613405] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267333, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.798143] env[62558]: DEBUG nova.compute.manager [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Received event network-changed-44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.798361] env[62558]: DEBUG nova.compute.manager [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Refreshing instance network info cache due to event network-changed-44bbeb25-dd9d-4cc1-914a-a3edad861b31. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.798532] env[62558]: DEBUG oslo_concurrency.lockutils [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.798717] env[62558]: DEBUG oslo_concurrency.lockutils [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.798893] env[62558]: DEBUG nova.network.neutron [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Refreshing network info cache for port 44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1054.011490] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267330, 'name': Rename_Task, 'duration_secs': 0.147} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.011818] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.012108] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15465a33-5dda-4885-b777-48bf355665bf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.023763] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1054.023763] env[62558]: value = "task-1267334" [ 1054.023763] env[62558]: _type = "Task" [ 1054.023763] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.032637] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.079965] env[62558]: DEBUG nova.objects.instance [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.117245] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267333, 'name': CreateVM_Task, 'duration_secs': 0.304082} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.117456] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.118178] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.118364] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.118701] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.118976] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52939525-3173-4a13-a19d-9b6089dfef73 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.126732] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1054.126732] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520adbcc-25b5-7f5d-4246-daf622bc5345" [ 1054.126732] env[62558]: _type = "Task" [ 1054.126732] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.136503] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520adbcc-25b5-7f5d-4246-daf622bc5345, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.333815] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.334056] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.509821] env[62558]: DEBUG nova.network.neutron [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updated VIF entry in instance network info cache for port 44bbeb25-dd9d-4cc1-914a-a3edad861b31. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1054.509821] env[62558]: DEBUG nova.network.neutron [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.536086] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267334, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.639347] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]520adbcc-25b5-7f5d-4246-daf622bc5345, 'name': SearchDatastore_Task, 'duration_secs': 0.014747} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.639791] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.640129] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.640399] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.640541] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.640775] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.641166] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05f3cad6-6896-4399-ad77-de67fa6e5365 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.651055] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.651248] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.652194] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8eb704d-4492-4f6f-b326-dd6366adf9dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.658010] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1054.658010] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5216a354-2602-98ab-70d7-e7e65e26ccdd" [ 1054.658010] env[62558]: _type = "Task" [ 1054.658010] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.667184] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5216a354-2602-98ab-70d7-e7e65e26ccdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.838082] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1055.012971] env[62558]: DEBUG oslo_concurrency.lockutils [req-40975f66-92a6-444e-9c9d-cde9972aff71 req-2f2a159b-8d6e-4d22-9285-60c96b3943f9 service nova] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.033266] env[62558]: DEBUG oslo_vmware.api [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267334, 'name': PowerOnVM_Task, 'duration_secs': 0.556141} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.034320] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.034530] env[62558]: INFO nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Took 8.96 seconds to spawn the instance on the hypervisor. [ 1055.034727] env[62558]: DEBUG nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.035591] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fadbac-1043-49a0-8b9b-36a599ed50e0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.088180] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5b9d5f10-9fd9-4208-99e4-86892a2ecffe tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.767s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.168958] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5216a354-2602-98ab-70d7-e7e65e26ccdd, 'name': SearchDatastore_Task, 'duration_secs': 0.033887} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.169753] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3f87303-ec89-45d2-aa50-1e68292f10c2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.175070] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1055.175070] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52739a6e-4493-3c94-8b9d-6c98364d92c9" [ 1055.175070] env[62558]: _type = "Task" [ 1055.175070] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.182435] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52739a6e-4493-3c94-8b9d-6c98364d92c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.359884] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.360161] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.361674] env[62558]: INFO nova.compute.claims [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.552820] env[62558]: INFO nova.compute.manager [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Took 16.99 seconds to build instance. [ 1055.654175] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.654447] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.684901] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52739a6e-4493-3c94-8b9d-6c98364d92c9, 'name': SearchDatastore_Task, 'duration_secs': 0.043807} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.685164] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.685464] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.685722] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc10ee1a-b9d0-42ef-bf3a-911c0d521225 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.692354] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1055.692354] env[62558]: value = "task-1267336" [ 1055.692354] env[62558]: _type = "Task" [ 1055.692354] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.699185] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267336, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.055192] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a239a908-1f44-4598-880e-067cdccea503 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.500s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.157973] env[62558]: INFO nova.compute.manager [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Detaching volume d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0 [ 1056.200627] env[62558]: INFO nova.virt.block_device [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Attempting to driver detach volume d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0 from mountpoint /dev/sdc [ 1056.200922] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1056.201166] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272616', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'name': 'volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'serial': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1056.202046] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ec7170-97df-4384-b72f-9b6d171151ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.209972] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267336, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493973} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.225344] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.226813] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.227225] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5d31b2d-6893-4efc-852d-6429ca6b72a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.233779] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff94104-8bdb-4e9f-9b1a-549ac647c137 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.244250] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6055201b-5843-42d6-8335-287336995c2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.244460] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1056.244460] env[62558]: value = "task-1267338" [ 1056.244460] env[62558]: _type = "Task" [ 1056.244460] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.263714] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92546b50-defc-4e8d-8b0d-703ad59491f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.268730] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267338, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.282150] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] The volume has not been displaced from its original location: [datastore2] volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0/volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1056.287388] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfiguring VM instance instance-00000062 to detach disk 2002 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1056.287723] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66ebd631-a326-43fd-b611-b65478f41c8c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.306014] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1056.306014] env[62558]: value = "task-1267339" [ 1056.306014] env[62558]: _type = "Task" [ 1056.306014] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.314133] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267339, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.469052] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba33d128-bf93-4c94-ad05-6e91bb79ce2b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.479159] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1519e2ec-a304-440f-9528-da54683d0562 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.512512] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdaeda41-4f89-4fb8-94e0-700540e5e0de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.520615] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5612d7ff-06c8-4ff5-8d0c-7913c425ef8f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.535466] env[62558]: DEBUG nova.compute.provider_tree [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1056.730553] env[62558]: DEBUG nova.compute.manager [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.731504] env[62558]: DEBUG nova.compute.manager [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1056.731504] env[62558]: DEBUG oslo_concurrency.lockutils [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.732013] env[62558]: DEBUG oslo_concurrency.lockutils [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.732330] env[62558]: DEBUG nova.network.neutron [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.757294] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267338, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.817841] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267339, 'name': ReconfigVM_Task, 'duration_secs': 0.310671} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.818356] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Reconfigured VM instance instance-00000062 to detach disk 2002 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1056.823055] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4d429e9-37ea-42a8-9ae3-321398d3d7ab {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.837915] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1056.837915] env[62558]: value = "task-1267340" [ 1056.837915] env[62558]: _type = "Task" [ 1056.837915] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.846288] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267340, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.055569] env[62558]: ERROR nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [req-f0ea181f-636b-4d61-a874-977c6c8d19f4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID dc830c09-1c36-446a-8af3-d3826bec8b3b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f0ea181f-636b-4d61-a874-977c6c8d19f4"}]} [ 1057.071394] env[62558]: DEBUG nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1057.090210] env[62558]: DEBUG nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1057.090210] env[62558]: DEBUG nova.compute.provider_tree [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1057.101355] env[62558]: DEBUG nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1057.128745] env[62558]: DEBUG nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1057.226417] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bcfaff-dd86-47da-b302-63afc4d7f4fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.234109] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3997718-a69a-43e0-b469-e0cc20fbb407 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.272256] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73cc2fe-ab22-4c61-95bf-85d6c5f79d9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.280236] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267338, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.283363] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4e5e8e-99ec-4ca6-a716-3d26580a3136 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.296653] env[62558]: DEBUG nova.compute.provider_tree [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1057.347988] env[62558]: DEBUG oslo_vmware.api [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267340, 'name': ReconfigVM_Task, 'duration_secs': 0.185344} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.348277] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272616', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'name': 'volume-d269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4e8b3008-693d-4445-937b-8e7db13c7c91', 'attached_at': '', 'detached_at': '', 'volume_id': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0', 'serial': 'd269c32e-c8df-4cfa-93cd-8b1f45ca5bb0'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1057.499805] env[62558]: DEBUG nova.network.neutron [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.500205] env[62558]: DEBUG nova.network.neutron [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.775696] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267338, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.082706} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.776027] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.776976] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1b69b3-1df9-4081-b293-2da9974d15d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.799695] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.803388] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-340a59c3-53c9-445e-a593-ae80ad1ab017 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.818537] env[62558]: DEBUG nova.compute.manager [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.819128] env[62558]: DEBUG nova.compute.manager [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing instance network info cache due to event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.819128] env[62558]: DEBUG oslo_concurrency.lockutils [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.819128] env[62558]: DEBUG oslo_concurrency.lockutils [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.819323] env[62558]: DEBUG nova.network.neutron [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.826963] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1057.826963] env[62558]: value = "task-1267341" [ 1057.826963] env[62558]: _type = "Task" [ 1057.826963] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.835850] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267341, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.844733] env[62558]: DEBUG nova.scheduler.client.report [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 135 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1057.844983] env[62558]: DEBUG nova.compute.provider_tree [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 135 to 136 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1057.845187] env[62558]: DEBUG nova.compute.provider_tree [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1057.895031] env[62558]: DEBUG nova.objects.instance [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'flavor' on Instance uuid 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.003544] env[62558]: DEBUG oslo_concurrency.lockutils [req-e0348b5f-0548-403f-8526-4dbc54b9ac83 req-4aefb038-d650-4cd3-9cd3-ce3e910fecf3 service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.337450] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267341, 'name': ReconfigVM_Task, 'duration_secs': 0.289276} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.337618] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.338324] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd3cd392-9dfc-4408-852a-ac1e5e0b87e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.345575] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1058.345575] env[62558]: value = "task-1267343" [ 1058.345575] env[62558]: _type = "Task" [ 1058.345575] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.349366] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.989s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.349860] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1058.357324] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267343, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.630750] env[62558]: DEBUG nova.network.neutron [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updated VIF entry in instance network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.631137] env[62558]: DEBUG nova.network.neutron [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.753985] env[62558]: DEBUG nova.compute.manager [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.754216] env[62558]: DEBUG nova.compute.manager [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing instance network info cache due to event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.754416] env[62558]: DEBUG oslo_concurrency.lockutils [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.855120] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267343, 'name': Rename_Task, 'duration_secs': 0.139609} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.855469] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.855728] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca4f8528-4064-425f-94da-103658390c17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.859680] env[62558]: DEBUG nova.compute.utils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.860988] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1058.861168] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1058.865884] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1058.865884] env[62558]: value = "task-1267344" [ 1058.865884] env[62558]: _type = "Task" [ 1058.865884] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.873510] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.901541] env[62558]: DEBUG nova.policy [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37bb028d611e42d98319c4afba5b1b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46aa5789fccf4925aa49f9c4050c8463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1058.903539] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e59ead22-6df0-4a7a-acd6-8afb6ffb5e92 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.249s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.135753] env[62558]: DEBUG oslo_concurrency.lockutils [req-a60e4d41-d62d-41b7-a651-69b6c7e75585 req-4b425982-816b-4ded-8e6a-0d173322befb service nova] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.136234] env[62558]: DEBUG oslo_concurrency.lockutils [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.136430] env[62558]: DEBUG nova.network.neutron [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.190769] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Successfully created port: 7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.366527] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1059.381300] env[62558]: DEBUG oslo_vmware.api [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267344, 'name': PowerOnVM_Task, 'duration_secs': 0.450389} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.381582] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.381785] env[62558]: INFO nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Took 8.56 seconds to spawn the instance on the hypervisor. [ 1059.381970] env[62558]: DEBUG nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.382851] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073f94d6-f88c-438a-8153-096d7d3945be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.824340] env[62558]: DEBUG nova.compute.manager [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.824508] env[62558]: DEBUG nova.compute.manager [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.824731] env[62558]: DEBUG oslo_concurrency.lockutils [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.824876] env[62558]: DEBUG oslo_concurrency.lockutils [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.825261] env[62558]: DEBUG nova.network.neutron [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.836239] env[62558]: DEBUG nova.network.neutron [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updated VIF entry in instance network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.836623] env[62558]: DEBUG nova.network.neutron [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.901134] env[62558]: INFO nova.compute.manager [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Took 16.44 seconds to build instance. [ 1060.049609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.049905] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.050137] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.050326] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.050502] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.052608] env[62558]: INFO nova.compute.manager [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Terminating instance [ 1060.054349] env[62558]: DEBUG nova.compute.manager [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1060.054543] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1060.055397] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fadcfde-6101-4391-949c-67f06f372d85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.063154] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.063392] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-263d9aad-d431-4f93-9fd1-1e3727f23b0f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.070282] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1060.070282] env[62558]: value = "task-1267345" [ 1060.070282] env[62558]: _type = "Task" [ 1060.070282] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.077980] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267345, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.339246] env[62558]: DEBUG oslo_concurrency.lockutils [req-f7497f40-1e62-42d7-8879-dbe4561d8d91 req-26725135-b69c-4244-b888-8bea5ad14a6e service nova] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.382016] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1060.402271] env[62558]: DEBUG oslo_concurrency.lockutils [None req-167bb41b-7ab0-40d7-842b-ec5bca29f880 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.947s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.412725] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.412970] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.413149] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.413334] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.413485] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.413679] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.413986] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.414433] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.414633] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.414811] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.414987] env[62558]: DEBUG nova.virt.hardware [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.416066] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67f4219-9cab-4ed3-ba88-e25036b7097b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.424246] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c92f29-4cae-4f4f-80cf-1f4fca6c2057 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.559153] env[62558]: DEBUG nova.network.neutron [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1060.559824] env[62558]: DEBUG nova.network.neutron [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.583517] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267345, 'name': PowerOffVM_Task, 'duration_secs': 0.259069} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.584115] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1060.584351] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1060.584715] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-458dcb88-0a70-4543-a8b8-b7c6d200e918 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.643974] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1060.644290] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1060.644393] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleting the datastore file [datastore2] 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1060.644670] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddf55bcc-31cd-45ce-b0dc-e40de26b17ea {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.652196] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for the task: (returnval){ [ 1060.652196] env[62558]: value = "task-1267348" [ 1060.652196] env[62558]: _type = "Task" [ 1060.652196] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.661116] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.701684] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Successfully updated port: 7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1060.785727] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Received event network-changed-44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.785947] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Refreshing instance network info cache due to event network-changed-44bbeb25-dd9d-4cc1-914a-a3edad861b31. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.786389] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.786648] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.786708] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Refreshing network info cache for port 44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1061.063510] env[62558]: DEBUG oslo_concurrency.lockutils [req-0048b77b-ffa7-4369-b96b-33e1d67e6b2a req-30690428-6aa3-4852-954f-490c446adf2b service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.162637] env[62558]: DEBUG oslo_vmware.api [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Task: {'id': task-1267348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158292} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.162937] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.163161] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.163354] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.163570] env[62558]: INFO nova.compute.manager [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1061.163853] env[62558]: DEBUG oslo.service.loopingcall [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.164104] env[62558]: DEBUG nova.compute.manager [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1061.164212] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.205085] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.205241] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.206180] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.504629] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updated VIF entry in instance network info cache for port 44bbeb25-dd9d-4cc1-914a-a3edad861b31. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.505146] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.963189] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1062.007477] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.007779] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Received event network-vif-plugged-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.007987] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.008251] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Lock "301dde86-496e-45b7-a86e-5bda587908db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.008445] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Lock "301dde86-496e-45b7-a86e-5bda587908db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.008645] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] No waiting events found dispatching network-vif-plugged-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.008839] env[62558]: WARNING nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Received unexpected event network-vif-plugged-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 for instance with vm_state building and task_state spawning. [ 1062.009044] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Received event network-changed-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.009242] env[62558]: DEBUG nova.compute.manager [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Refreshing instance network info cache due to event network-changed-7c11b24d-6b71-48d9-ae68-a7c4f88c7367. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.009443] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Acquiring lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.085877] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.115451] env[62558]: DEBUG nova.network.neutron [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating instance_info_cache with network_info: [{"id": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "address": "fa:16:3e:61:71:40", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c11b24d-6b", "ovs_interfaceid": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.590083] env[62558]: INFO nova.compute.manager [-] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Took 1.42 seconds to deallocate network for instance. [ 1062.618688] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.619041] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Instance network_info: |[{"id": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "address": "fa:16:3e:61:71:40", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c11b24d-6b", "ovs_interfaceid": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1062.619436] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Acquired lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.619653] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Refreshing network info cache for port 7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.620948] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:71:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7c11b24d-6b71-48d9-ae68-a7c4f88c7367', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.628898] env[62558]: DEBUG oslo.service.loopingcall [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.632265] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.632770] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76ba792f-2226-41ea-9c30-5bc8b3572108 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.652746] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.652746] env[62558]: value = "task-1267350" [ 1062.652746] env[62558]: _type = "Task" [ 1062.652746] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.661772] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267350, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.812403] env[62558]: DEBUG nova.compute.manager [req-8cb3d005-7eef-4304-89b8-2716b192a41d req-c026b780-e5de-4714-9537-b459c6b8e8dc service nova] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Received event network-vif-deleted-c8e25a3a-b0f5-428b-a485-ff34df8b7fe5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.846585] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updated VIF entry in instance network info cache for port 7c11b24d-6b71-48d9-ae68-a7c4f88c7367. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1062.847069] env[62558]: DEBUG nova.network.neutron [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating instance_info_cache with network_info: [{"id": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "address": "fa:16:3e:61:71:40", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c11b24d-6b", "ovs_interfaceid": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.096925] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.097291] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.097436] env[62558]: DEBUG nova.objects.instance [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lazy-loading 'resources' on Instance uuid 4e8b3008-693d-4445-937b-8e7db13c7c91 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.162957] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267350, 'name': CreateVM_Task, 'duration_secs': 0.341708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.163140] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.163797] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.163957] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.164295] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.164539] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-443f1291-dc4b-4470-9d3b-8906f6602830 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.168650] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1063.168650] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f32723-c352-308a-d254-70f86aa85073" [ 1063.168650] env[62558]: _type = "Task" [ 1063.168650] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.176098] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f32723-c352-308a-d254-70f86aa85073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.350377] env[62558]: DEBUG oslo_concurrency.lockutils [req-8a9d6372-40d5-43f4-8e17-5e6623b0eb91 req-2c64479e-6721-44b2-9f75-246647d375cc service nova] Releasing lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.680571] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52f32723-c352-308a-d254-70f86aa85073, 'name': SearchDatastore_Task, 'duration_secs': 0.009734} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.680880] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.681132] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.681367] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.681603] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.681691] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.683925] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86ca614d-829d-4b8f-876c-defed135d4bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.691933] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.692127] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1063.692838] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4772bbed-1c1c-4551-b51c-b98ce3378d51 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.696098] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48aa034d-c106-4d7f-87ef-4f2687130cc2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.701334] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1063.701334] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bf9f9c-66d4-f71e-a26e-de6abdce4709" [ 1063.701334] env[62558]: _type = "Task" [ 1063.701334] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.706200] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52877ff-219d-4bf8-a486-8e4f7730706b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.714384] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bf9f9c-66d4-f71e-a26e-de6abdce4709, 'name': SearchDatastore_Task, 'duration_secs': 0.008685} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.737948] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e14678-c280-49a2-b3bf-f2ae9b5e3a58 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.740549] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c242c0-dfc9-4ff6-bfff-582e5826cb03 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.746806] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1063.746806] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c8ff2f-e41e-d6bb-17b8-83be58c35df7" [ 1063.746806] env[62558]: _type = "Task" [ 1063.746806] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.756521] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4f8190-ddbe-486e-bc6b-2f751e9f2254 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.765693] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c8ff2f-e41e-d6bb-17b8-83be58c35df7, 'name': SearchDatastore_Task, 'duration_secs': 0.014431} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.773021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.773305] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 301dde86-496e-45b7-a86e-5bda587908db/301dde86-496e-45b7-a86e-5bda587908db.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1063.773780] env[62558]: DEBUG nova.compute.provider_tree [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.775058] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04f8ed23-61b1-405d-9ac7-58f29e6ea114 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.781606] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1063.781606] env[62558]: value = "task-1267351" [ 1063.781606] env[62558]: _type = "Task" [ 1063.781606] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.790477] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.293164] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267351, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457634} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.293477] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 301dde86-496e-45b7-a86e-5bda587908db/301dde86-496e-45b7-a86e-5bda587908db.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1064.293646] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.293912] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8063d832-a92a-4e47-a305-715e05899dc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.300384] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1064.300384] env[62558]: value = "task-1267352" [ 1064.300384] env[62558]: _type = "Task" [ 1064.300384] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.308187] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267352, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.309094] env[62558]: DEBUG nova.scheduler.client.report [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Updated inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1064.309336] env[62558]: DEBUG nova.compute.provider_tree [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Updating resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b generation from 136 to 137 during operation: update_inventory {{(pid=62558) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1064.309522] env[62558]: DEBUG nova.compute.provider_tree [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1064.630684] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.630908] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.631037] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1064.631235] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.631365] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1064.810940] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267352, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060897} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.811244] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.812070] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf089487-f9f6-454f-b916-0516d99347a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.815121] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.837086] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 301dde86-496e-45b7-a86e-5bda587908db/301dde86-496e-45b7-a86e-5bda587908db.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.838137] env[62558]: INFO nova.scheduler.client.report [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Deleted allocations for instance 4e8b3008-693d-4445-937b-8e7db13c7c91 [ 1064.839068] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34860120-713b-415b-9744-3e13685863e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.860399] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1064.860399] env[62558]: value = "task-1267354" [ 1064.860399] env[62558]: _type = "Task" [ 1064.860399] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.868601] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267354, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.153361] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] There are 51 instances to clean {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1065.153833] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: e7569632-751d-4bd7-b00a-5176c816551c] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.358866] env[62558]: DEBUG oslo_concurrency.lockutils [None req-aa571e7d-dc2d-4ad4-942a-c38d34164eb1 tempest-AttachVolumeTestJSON-1181789984 tempest-AttachVolumeTestJSON-1181789984-project-member] Lock "4e8b3008-693d-4445-937b-8e7db13c7c91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.309s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.370108] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267354, 'name': ReconfigVM_Task, 'duration_secs': 0.260698} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.370752] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 301dde86-496e-45b7-a86e-5bda587908db/301dde86-496e-45b7-a86e-5bda587908db.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.371016] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c2e5e0a-1d7e-4a6e-9f56-1548dda518dd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.377569] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1065.377569] env[62558]: value = "task-1267355" [ 1065.377569] env[62558]: _type = "Task" [ 1065.377569] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.385487] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267355, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.658804] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4dffea0b-adb7-4c6a-b2b0-436796111958] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.886585] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267355, 'name': Rename_Task, 'duration_secs': 0.1468} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.886907] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.887177] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65d9682b-ad01-42ce-8304-a8c95e778c3b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.893331] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1065.893331] env[62558]: value = "task-1267357" [ 1065.893331] env[62558]: _type = "Task" [ 1065.893331] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.900769] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267357, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.163115] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: a94f43ba-da37-479c-b939-016dff83c754] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.403463] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267357, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.666109] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6687a008-3193-4a43-b67c-c10e5ffcfc11] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.904108] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267357, 'name': PowerOnVM_Task} progress is 79%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.169502] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9b87ed06-b062-4e6f-9dfa-a9b5491fe101] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.407099] env[62558]: DEBUG oslo_vmware.api [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267357, 'name': PowerOnVM_Task, 'duration_secs': 1.160615} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.407449] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.407653] env[62558]: INFO nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Took 7.03 seconds to spawn the instance on the hypervisor. [ 1067.407890] env[62558]: DEBUG nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.408914] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7faa42-5fa6-496e-9902-cfda4d3ff9ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.672948] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ea680004-b94a-47f9-b516-2afb95c30a51] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.929861] env[62558]: INFO nova.compute.manager [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Took 12.59 seconds to build instance. [ 1068.176688] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ec2b48ff-dffa-4948-9925-a1c23aa76e64] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.432436] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4aa231e8-262d-4474-9aeb-4cd0585ff74e tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.098s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.661714] env[62558]: DEBUG nova.compute.manager [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Received event network-changed-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.661714] env[62558]: DEBUG nova.compute.manager [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Refreshing instance network info cache due to event network-changed-7c11b24d-6b71-48d9-ae68-a7c4f88c7367. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1068.661714] env[62558]: DEBUG oslo_concurrency.lockutils [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] Acquiring lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.661839] env[62558]: DEBUG oslo_concurrency.lockutils [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] Acquired lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.661956] env[62558]: DEBUG nova.network.neutron [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Refreshing network info cache for port 7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1068.679954] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ef0b3d16-704d-4435-9c23-a258d94a9983] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.182667] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 591d326c-c20c-401b-a57b-895d59c3b418] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.387730] env[62558]: DEBUG nova.network.neutron [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updated VIF entry in instance network info cache for port 7c11b24d-6b71-48d9-ae68-a7c4f88c7367. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1069.388244] env[62558]: DEBUG nova.network.neutron [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating instance_info_cache with network_info: [{"id": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "address": "fa:16:3e:61:71:40", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c11b24d-6b", "ovs_interfaceid": "7c11b24d-6b71-48d9-ae68-a7c4f88c7367", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.608174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.608609] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.685554] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: c65dbe09-d073-48ad-a18a-8b6383c7e345] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.891250] env[62558]: DEBUG oslo_concurrency.lockutils [req-f3100946-8f18-4e53-943b-f0f86dd0559c req-cf9bfd91-a2af-4232-8581-4be11e6455dd service nova] Releasing lock "refresh_cache-301dde86-496e-45b7-a86e-5bda587908db" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.111180] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1070.188834] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 109f67db-21cf-40f9-966e-4730bff98bcf] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.632543] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.632812] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.634387] env[62558]: INFO nova.compute.claims [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.692458] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 7188fa81-3a23-4f22-adb8-41a35b1af9a5] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.195160] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ab4d0062-2ae0-4a34-9f6d-1cd2dc1e0568] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.699247] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 29584082-37e6-4dd8-906f-fd87b4ca9bc2] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.738369] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de8ae1d-8eee-4e29-8f61-cb1d3f6600c7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.747539] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c69272d-686c-4819-be1a-0391b6bae5a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.778221] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e31b19-5502-4c96-83f5-54b382333d4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.786334] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dba8305-59fe-425a-a1ba-d4fe2a7be22f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.801434] env[62558]: DEBUG nova.compute.provider_tree [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.100123] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.100336] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.100565] env[62558]: DEBUG nova.objects.instance [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid 5f75bf07-002f-4ed4-930a-fbe6f939cb4f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.202606] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2480822a-e10a-4066-a5d8-5ca633ab9b12] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.305321] env[62558]: DEBUG nova.scheduler.client.report [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.696984] env[62558]: DEBUG nova.objects.instance [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid 5f75bf07-002f-4ed4-930a-fbe6f939cb4f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.705766] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 42b2e70f-a46b-4984-b6d1-a2bfe348d1f6] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.810924] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.811359] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1073.199983] env[62558]: DEBUG nova.objects.base [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance<5f75bf07-002f-4ed4-930a-fbe6f939cb4f> lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.200168] env[62558]: DEBUG nova.network.neutron [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1073.209120] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f8d6df86-52b8-4692-add4-1ffec82cc598] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.267195] env[62558]: DEBUG nova.policy [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1073.316165] env[62558]: DEBUG nova.compute.utils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.317588] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1073.317754] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1073.356032] env[62558]: DEBUG nova.policy [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4e6e2f1bae4d4bb68070770dbbbe5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78b69738b90142a4943069006e349445', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1073.626035] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Successfully created port: f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1073.712895] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: a3f42034-3d2b-472a-89c0-5445cb6fb567] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.820882] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1074.215772] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f642911c-bb9e-4187-8e01-65c45cb6e793] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.326959] env[62558]: INFO nova.virt.block_device [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Booting with volume 393854fc-2de5-43ff-8f13-b76ec330d217 at /dev/sda [ 1074.366231] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-648f634f-f502-4cd6-aea2-7fc58de27714 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.377632] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a20a4c6-fe1f-44a6-9044-0e9f430859a7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.404940] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9119e17f-d28b-4893-a2c2-962645c8b65f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.413118] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0ec979-d126-4e54-a4d2-9ae60857b35e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.442482] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a2d141-f134-4e50-9d12-6b44c63745c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.449580] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac11e4f6-595e-46cb-872d-4f613f770003 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.468087] env[62558]: DEBUG nova.virt.block_device [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating existing volume attachment record: bc6a68d2-3129-4a34-91ec-f26a08b9b777 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1074.721058] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ecf01ab6-5019-4453-b102-0e754abc4ef8] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.774294] env[62558]: DEBUG nova.compute.manager [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-vif-plugged-24332129-4159-4273-80de-84706acf064a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.774534] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.775316] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.775582] env[62558]: DEBUG oslo_concurrency.lockutils [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.777877] env[62558]: DEBUG nova.compute.manager [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] No waiting events found dispatching network-vif-plugged-24332129-4159-4273-80de-84706acf064a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1074.778191] env[62558]: WARNING nova.compute.manager [req-9d19eeba-7942-45af-b343-0a30d9d60e72 req-21325a60-d458-4023-90f9-956766096e21 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received unexpected event network-vif-plugged-24332129-4159-4273-80de-84706acf064a for instance with vm_state active and task_state None. [ 1074.874176] env[62558]: DEBUG nova.network.neutron [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Successfully updated port: 24332129-4159-4273-80de-84706acf064a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.135912] env[62558]: DEBUG nova.compute.manager [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Received event network-vif-plugged-f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.136170] env[62558]: DEBUG oslo_concurrency.lockutils [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.136385] env[62558]: DEBUG oslo_concurrency.lockutils [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.136555] env[62558]: DEBUG oslo_concurrency.lockutils [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.136725] env[62558]: DEBUG nova.compute.manager [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] No waiting events found dispatching network-vif-plugged-f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1075.136944] env[62558]: WARNING nova.compute.manager [req-c9ee8ce3-4463-42ae-ab33-76a085563e29 req-3ba1abca-7d87-4b55-a6c3-7bb88dbedb7d service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Received unexpected event network-vif-plugged-f7bda9aa-e43a-4d03-a623-e549e9529118 for instance with vm_state building and task_state block_device_mapping. [ 1075.224651] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 7aebdb8a-517b-4168-91e0-1b704e6a11c7] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.308261] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Successfully updated port: f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.377425] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.377695] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.377822] env[62558]: DEBUG nova.network.neutron [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.728326] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 016c31f6-648c-41f5-909f-5a1ae6366b98] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.810436] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.810587] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.810735] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1075.920038] env[62558]: WARNING nova.network.neutron [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 1076.232118] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6c2fedb1-6240-482d-9ad2-e8dabde4d7ed] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.258880] env[62558]: DEBUG nova.network.neutron [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24332129-4159-4273-80de-84706acf064a", "address": "fa:16:3e:ba:fc:4a", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24332129-41", "ovs_interfaceid": "24332129-4159-4273-80de-84706acf064a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.338714] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1076.459119] env[62558]: DEBUG nova.network.neutron [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.563844] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1076.564756] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.565123] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.565409] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.565678] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.565875] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.566046] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.566266] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.566432] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.566605] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.566775] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.566987] env[62558]: DEBUG nova.virt.hardware [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.567837] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f431f980-9c9a-4f74-9cab-bce68d342a65 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.576075] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a71b544-b6a9-4839-a01c-ba188f4ba22f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.735484] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 0e5c1138-7f09-4b76-9c56-93c0ad947ad8] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.761830] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.762524] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.762714] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.763857] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22744434-92ee-47b5-bba2-149563297180 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.780618] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.780847] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.781015] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.781208] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.781357] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.781504] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.781704] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.781864] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.782050] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.782224] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.782400] env[62558]: DEBUG nova.virt.hardware [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.788629] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfiguring VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1076.788905] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5d2d925-04bd-44b0-8572-ffc2462a2a0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.802477] env[62558]: DEBUG nova.compute.manager [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-24332129-4159-4273-80de-84706acf064a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1076.802665] env[62558]: DEBUG nova.compute.manager [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-24332129-4159-4273-80de-84706acf064a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1076.802870] env[62558]: DEBUG oslo_concurrency.lockutils [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.803027] env[62558]: DEBUG oslo_concurrency.lockutils [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.803190] env[62558]: DEBUG nova.network.neutron [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port 24332129-4159-4273-80de-84706acf064a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1076.809055] env[62558]: DEBUG oslo_vmware.api [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1076.809055] env[62558]: value = "task-1267361" [ 1076.809055] env[62558]: _type = "Task" [ 1076.809055] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.818274] env[62558]: DEBUG oslo_vmware.api [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267361, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.961530] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.961909] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Instance network_info: |[{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1076.962377] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:bb:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ff81f9-72b2-4e58-a8d8-5699907f7459', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7bda9aa-e43a-4d03-a623-e549e9529118', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.969963] env[62558]: DEBUG oslo.service.loopingcall [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.970243] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.970879] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c01bca44-733c-4793-b27f-ab5df0319f06 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.991078] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.991078] env[62558]: value = "task-1267362" [ 1076.991078] env[62558]: _type = "Task" [ 1076.991078] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.998710] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267362, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.168305] env[62558]: DEBUG nova.compute.manager [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Received event network-changed-f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.168376] env[62558]: DEBUG nova.compute.manager [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Refreshing instance network info cache due to event network-changed-f7bda9aa-e43a-4d03-a623-e549e9529118. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1077.168672] env[62558]: DEBUG oslo_concurrency.lockutils [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.168799] env[62558]: DEBUG oslo_concurrency.lockutils [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.168890] env[62558]: DEBUG nova.network.neutron [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Refreshing network info cache for port f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.239289] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 0aeb55df-48e8-4314-ab47-05ff0b5eaa7b] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.318346] env[62558]: DEBUG oslo_vmware.api [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.501245] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267362, 'name': CreateVM_Task, 'duration_secs': 0.308026} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.501426] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.502114] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': 'bc6a68d2-3129-4a34-91ec-f26a08b9b777', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272623', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'name': 'volume-393854fc-2de5-43ff-8f13-b76ec330d217', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be1007a-6229-4d45-82d0-53d81a28204c', 'attached_at': '', 'detached_at': '', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'serial': '393854fc-2de5-43ff-8f13-b76ec330d217'}, 'boot_index': 0, 'disk_bus': None, 'delete_on_termination': True, 'guest_format': None, 'mount_device': '/dev/sda', 'volume_type': None}], 'swap': None} {{(pid=62558) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1077.502470] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Root volume attach. Driver type: vmdk {{(pid=62558) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1077.503129] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff122ae0-dc2d-4047-87fa-65f79000241c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.510882] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1e53c4-b83e-4332-844e-bdc93de28e4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.516891] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3eef83-20e1-4aaa-b533-c3e5668a1059 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.522433] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e9fc30f2-8751-48a0-a36f-9a23d7b2b1e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.529014] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1077.529014] env[62558]: value = "task-1267363" [ 1077.529014] env[62558]: _type = "Task" [ 1077.529014] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.536154] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267363, 'name': RelocateVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.680333] env[62558]: DEBUG nova.network.neutron [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port 24332129-4159-4273-80de-84706acf064a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1077.681158] env[62558]: DEBUG nova.network.neutron [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24332129-4159-4273-80de-84706acf064a", "address": "fa:16:3e:ba:fc:4a", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24332129-41", "ovs_interfaceid": "24332129-4159-4273-80de-84706acf064a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.742635] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4815ba3f-265f-466a-9850-4c325cdb88de] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.820020] env[62558]: DEBUG oslo_vmware.api [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267361, 'name': ReconfigVM_Task, 'duration_secs': 0.600484} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.820502] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.820715] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfigured VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1077.890688] env[62558]: DEBUG nova.network.neutron [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updated VIF entry in instance network info cache for port f7bda9aa-e43a-4d03-a623-e549e9529118. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1077.891069] env[62558]: DEBUG nova.network.neutron [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.038854] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267363, 'name': RelocateVM_Task, 'duration_secs': 0.025944} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.041032] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1078.041032] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272623', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'name': 'volume-393854fc-2de5-43ff-8f13-b76ec330d217', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be1007a-6229-4d45-82d0-53d81a28204c', 'attached_at': '', 'detached_at': '', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'serial': '393854fc-2de5-43ff-8f13-b76ec330d217'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1078.041032] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf30bf16-3da9-47b3-b536-4177a554a83d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.055347] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e4ed39-813b-47cc-be53-689f377b6855 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.076087] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-393854fc-2de5-43ff-8f13-b76ec330d217/volume-393854fc-2de5-43ff-8f13-b76ec330d217.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.076324] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20e693b7-69bf-4fea-a5e9-a8e385f7cdd3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.094461] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1078.094461] env[62558]: value = "task-1267364" [ 1078.094461] env[62558]: _type = "Task" [ 1078.094461] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.105479] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267364, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.182965] env[62558]: DEBUG oslo_concurrency.lockutils [req-1f1e2a4d-d4a4-4f1f-837a-8168a1857d8d req-0de2b51b-c17d-4d8e-a294-dd0c01ed5994 service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.248187] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 63050323-047e-4d73-91ae-859467b4b5a7] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.325493] env[62558]: DEBUG oslo_concurrency.lockutils [None req-68fb68ce-47bb-4b01-83b7-dd428e1aa370 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.225s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.393595] env[62558]: DEBUG oslo_concurrency.lockutils [req-c5b7f505-2723-4e4e-b5ad-7885b982e151 req-8568fc95-1483-4f16-ba08-bf10acb64aec service nova] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.605521] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267364, 'name': ReconfigVM_Task, 'duration_secs': 0.265657} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.605810] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-393854fc-2de5-43ff-8f13-b76ec330d217/volume-393854fc-2de5-43ff-8f13-b76ec330d217.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.610353] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1aba1c27-2db1-40ad-9348-d0ce9b0ffd6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.624661] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1078.624661] env[62558]: value = "task-1267365" [ 1078.624661] env[62558]: _type = "Task" [ 1078.624661] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.632120] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267365, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.751279] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 574525eb-0535-4664-8449-813c16e4781c] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.134643] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267365, 'name': ReconfigVM_Task, 'duration_secs': 0.116748} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.134947] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272623', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'name': 'volume-393854fc-2de5-43ff-8f13-b76ec330d217', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be1007a-6229-4d45-82d0-53d81a28204c', 'attached_at': '', 'detached_at': '', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'serial': '393854fc-2de5-43ff-8f13-b76ec330d217'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1079.135594] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ba8e4ec-ac96-4848-9a95-f6eaad204135 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.141030] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1079.141030] env[62558]: value = "task-1267366" [ 1079.141030] env[62558]: _type = "Task" [ 1079.141030] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.148060] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267366, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.254375] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2b8430ef-c8eb-4eb8-a754-3c552662b966] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.650593] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267366, 'name': Rename_Task, 'duration_secs': 0.122586} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.650888] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1079.651173] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c6dc6c2-3634-4020-9468-7b3f899b091d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.657585] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1079.657585] env[62558]: value = "task-1267367" [ 1079.657585] env[62558]: _type = "Task" [ 1079.657585] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.664817] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.757601] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 80221843-4e15-4f20-aeb4-4e6081371b95] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.769218] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.769499] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.167913] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267367, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.260747] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: a5d08f9b-ec56-4ad4-a4f5-f76bcb43f884] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.272653] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.272998] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.273905] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564b1902-fe4e-4dd3-ba57-1f3adc3f8835 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.292903] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40777d27-b91e-480f-961d-7f5468fd7849 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.318563] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfiguring VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1080.319056] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6391721-1c17-4c01-a765-3c6d3c9be580 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.337249] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1080.337249] env[62558]: value = "task-1267368" [ 1080.337249] env[62558]: _type = "Task" [ 1080.337249] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.346754] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.668037] env[62558]: DEBUG oslo_vmware.api [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267367, 'name': PowerOnVM_Task, 'duration_secs': 0.956906} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.668334] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1080.668586] env[62558]: INFO nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Took 4.10 seconds to spawn the instance on the hypervisor. [ 1080.668806] env[62558]: DEBUG nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1080.669593] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ed45f2-8355-4056-a8d6-a0a4dd12e731 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.763609] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2a7b4e08-d9b1-49f1-9611-ca1acc3315ca] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.847215] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.186202] env[62558]: INFO nova.compute.manager [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Took 10.57 seconds to build instance. [ 1081.266577] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6243631f-b240-4d7c-8910-0bf3d2dedb77] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.348300] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.688537] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f65fbaf6-5b42-4481-a44b-e47cbe001d8f tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.080s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.769336] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 66ff3b25-d49c-4129-8b51-5338e75b09b1] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.849918] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.920748] env[62558]: DEBUG nova.compute.manager [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.920958] env[62558]: DEBUG nova.compute.manager [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing instance network info cache due to event network-changed-6fcb24d4-f48e-4628-995d-10f53f01154b. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1081.921184] env[62558]: DEBUG oslo_concurrency.lockutils [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] Acquiring lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.921329] env[62558]: DEBUG oslo_concurrency.lockutils [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] Acquired lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.921488] env[62558]: DEBUG nova.network.neutron [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Refreshing network info cache for port 6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.273024] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 49a58b46-207f-4515-b313-afcdb2d1ced3] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.349348] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.605321] env[62558]: DEBUG nova.network.neutron [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updated VIF entry in instance network info cache for port 6fcb24d4-f48e-4628-995d-10f53f01154b. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1082.605699] env[62558]: DEBUG nova.network.neutron [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [{"id": "6fcb24d4-f48e-4628-995d-10f53f01154b", "address": "fa:16:3e:09:d7:9e", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb24d4-f4", "ovs_interfaceid": "6fcb24d4-f48e-4628-995d-10f53f01154b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.776799] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4526df2a-f801-4de8-8218-497a7d22034f] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.849062] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.108774] env[62558]: DEBUG oslo_concurrency.lockutils [req-8e550574-8fbb-4d20-ba18-8f3b573d1397 req-9d39ab2c-6781-46a3-a4d2-a7f777269662 service nova] Releasing lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.280370] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: e846e43d-68ec-4de3-ba62-e538643b6e4b] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.296443] env[62558]: DEBUG nova.compute.manager [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1083.350663] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.783772] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4a3f7642-1b9c-40d8-973b-5153b559bda0] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.850109] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.884885] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.885183] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.951196] env[62558]: DEBUG nova.compute.manager [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Received event network-changed-f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.951196] env[62558]: DEBUG nova.compute.manager [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Refreshing instance network info cache due to event network-changed-f7bda9aa-e43a-4d03-a623-e549e9529118. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1083.951365] env[62558]: DEBUG oslo_concurrency.lockutils [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.951515] env[62558]: DEBUG oslo_concurrency.lockutils [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.951677] env[62558]: DEBUG nova.network.neutron [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Refreshing network info cache for port f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.287301] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 8e3fa0dd-5f34-4a1d-8f94-b006b626f2b5] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.351748] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.391486] env[62558]: INFO nova.compute.claims [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1084.651802] env[62558]: DEBUG nova.network.neutron [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updated VIF entry in instance network info cache for port f7bda9aa-e43a-4d03-a623-e549e9529118. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1084.652146] env[62558]: DEBUG nova.network.neutron [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.790602] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ebd84b3f-0c5d-40e3-aa70-f8b3054b7109] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.851654] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.895605] env[62558]: INFO nova.compute.resource_tracker [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating resource usage from migration 9fd47a8b-ae6a-4ecf-b664-2f191a071d7f [ 1084.996607] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a2807c-a7c5-4abb-87e9-43d0dc2f9739 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.004468] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e3c3f0-deeb-4563-96f4-f8431b5c89bd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.032920] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55211138-c15b-441c-833a-f8160a675794 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.039577] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbc31f2-9c41-4401-9b21-b1a36abfbe76 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.053271] env[62558]: DEBUG nova.compute.provider_tree [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.154691] env[62558]: DEBUG oslo_concurrency.lockutils [req-fba220a0-50ac-4263-ab0c-0638dd3c9130 req-ea3989a9-b22e-41d9-bcef-97b476142755 service nova] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.293980] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ec8fdf55-07ad-4ec9-b913-a0a9b8c95c55] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.351664] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.557041] env[62558]: DEBUG nova.scheduler.client.report [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.797640] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: a1d242e6-1561-4bd4-8e39-281ab6346661] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.852802] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.063569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.178s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.063787] env[62558]: INFO nova.compute.manager [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Migrating [ 1086.300994] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6d89107f-0727-4b8c-863d-d1e635000bff] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.352772] env[62558]: DEBUG oslo_vmware.api [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267368, 'name': ReconfigVM_Task, 'duration_secs': 5.759744} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.353011] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.353204] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Reconfigured VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1086.579582] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.580657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.580657] env[62558]: DEBUG nova.network.neutron [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1086.804533] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: a8fef137-a651-4724-8ad1-a3e6ddcb5b03] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.288973] env[62558]: DEBUG nova.network.neutron [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.308102] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 3c8c2d25-981e-49de-b5c0-e0d12dd91378] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.693253] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.693504] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.693736] env[62558]: DEBUG nova.network.neutron [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.793268] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.812536] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 8acd06fc-b040-468c-980f-6e17b2343c4a] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.314092] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 99bcd698-cd39-4ed4-8c4d-54aa7a3cb7d1] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.556230] env[62558]: INFO nova.network.neutron [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Port 24332129-4159-4273-80de-84706acf064a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1088.556623] env[62558]: DEBUG nova.network.neutron [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.724111] env[62558]: DEBUG nova.compute.manager [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1088.724344] env[62558]: DEBUG nova.compute.manager [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing instance network info cache due to event network-changed-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1088.724541] env[62558]: DEBUG oslo_concurrency.lockutils [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] Acquiring lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.817560] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 2ac801d7-af70-46e0-88b3-02caee13497d] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.059529] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.061737] env[62558]: DEBUG oslo_concurrency.lockutils [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] Acquired lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.061867] env[62558]: DEBUG nova.network.neutron [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Refreshing network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.306499] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818445a5-4994-4f06-879c-d31ac0394714 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.325337] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: b0038711-5d12-4909-b331-72acb5ed0d24] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.327115] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.407527] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.407790] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.408163] env[62558]: DEBUG nova.objects.instance [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'flavor' on Instance uuid ac1ec03d-e04b-42eb-a1f3-c14887241bfa {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.567617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a833c01a-d070-4ec3-b171-1f29c3cde019 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-5f75bf07-002f-4ed4-930a-fbe6f939cb4f-24332129-4159-4273-80de-84706acf064a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.798s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.753602] env[62558]: DEBUG nova.network.neutron [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updated VIF entry in instance network info cache for port dedfab3d-cdf6-40b7-953a-7d13dad8f5a6. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1089.754070] env[62558]: DEBUG nova.network.neutron [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [{"id": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "address": "fa:16:3e:de:7e:80", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedfab3d-cd", "ovs_interfaceid": "dedfab3d-cdf6-40b7-953a-7d13dad8f5a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.833669] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1089.834058] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 6e01c855-ef25-45c9-a73e-5656e8e05cd4] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.835691] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2054387b-4782-4674-b91d-c56c1b0c334d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.843643] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1089.843643] env[62558]: value = "task-1267369" [ 1089.843643] env[62558]: _type = "Task" [ 1089.843643] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.853199] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.006213] env[62558]: DEBUG nova.objects.instance [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'pci_requests' on Instance uuid ac1ec03d-e04b-42eb-a1f3-c14887241bfa {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.257444] env[62558]: DEBUG oslo_concurrency.lockutils [req-6f90819b-ab23-42eb-8f86-e4db9605a4bb req-7de24d4e-57dc-47be-84bb-431d47ac406b service nova] Releasing lock "refresh_cache-5f75bf07-002f-4ed4-930a-fbe6f939cb4f" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.338859] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 867c0fb3-d4f6-4945-a5cc-6d93efcadb79] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.354392] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267369, 'name': PowerOffVM_Task, 'duration_secs': 0.179699} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.354392] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.354392] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.509639] env[62558]: DEBUG nova.objects.base [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1090.509871] env[62558]: DEBUG nova.network.neutron [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1090.569856] env[62558]: DEBUG nova.policy [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2a875c0ca0c476c905b86b12f5592e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd56f6c392f14aa880ef4891990c1d44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.748115] env[62558]: DEBUG nova.compute.manager [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.748401] env[62558]: DEBUG nova.compute.manager [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing instance network info cache due to event network-changed-66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1090.748561] env[62558]: DEBUG oslo_concurrency.lockutils [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.748681] env[62558]: DEBUG oslo_concurrency.lockutils [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.748842] env[62558]: DEBUG nova.network.neutron [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1090.841914] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1090.842190] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances with incomplete migration {{(pid=62558) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1090.859137] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.859398] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.859555] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.859736] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.859883] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.860040] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.860252] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.860417] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.860583] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.860747] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.860918] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.866140] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-645eb573-e11c-4f9f-bec5-e2f74ad02540 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.883444] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1090.883444] env[62558]: value = "task-1267370" [ 1090.883444] env[62558]: _type = "Task" [ 1090.883444] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.890751] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.344360] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.393299] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267370, 'name': ReconfigVM_Task, 'duration_secs': 0.173012} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.393566] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.447411] env[62558]: DEBUG nova.network.neutron [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updated VIF entry in instance network info cache for port 66f313e4-07aa-4e57-b472-a4408da24e26. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1091.447754] env[62558]: DEBUG nova.network.neutron [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.899707] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.900035] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.900171] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.900377] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.900501] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.900740] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.900973] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.901158] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.901331] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.901497] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.901670] env[62558]: DEBUG nova.virt.hardware [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.907776] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1091.908086] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38b70058-bdfc-459a-8399-5b5c7c483131 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.926118] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1091.926118] env[62558]: value = "task-1267371" [ 1091.926118] env[62558]: _type = "Task" [ 1091.926118] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.935874] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.951415] env[62558]: DEBUG oslo_concurrency.lockutils [req-e35aaef0-186a-445c-a3cc-23ba78f85b53 req-3c3a8d1a-4946-4edc-801d-1039456be730 service nova] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.006620] env[62558]: DEBUG nova.network.neutron [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Successfully updated port: 24332129-4159-4273-80de-84706acf064a {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.436791] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267371, 'name': ReconfigVM_Task, 'duration_secs': 0.149338} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.437104] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.437899] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202f0a02-cce8-4653-9791-bc1e6aabd67f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.459324] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-393854fc-2de5-43ff-8f13-b76ec330d217/volume-393854fc-2de5-43ff-8f13-b76ec330d217.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.459573] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6d6900c-5ff3-4121-9522-a9c58e42d0c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.477231] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1092.477231] env[62558]: value = "task-1267372" [ 1092.477231] env[62558]: _type = "Task" [ 1092.477231] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.484457] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.509490] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.509698] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.509912] env[62558]: DEBUG nova.network.neutron [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.772888] env[62558]: DEBUG nova.compute.manager [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-vif-plugged-24332129-4159-4273-80de-84706acf064a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.773139] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.773352] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.773524] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.773696] env[62558]: DEBUG nova.compute.manager [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] No waiting events found dispatching network-vif-plugged-24332129-4159-4273-80de-84706acf064a {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1092.773865] env[62558]: WARNING nova.compute.manager [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received unexpected event network-vif-plugged-24332129-4159-4273-80de-84706acf064a for instance with vm_state active and task_state None. [ 1092.774060] env[62558]: DEBUG nova.compute.manager [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-changed-24332129-4159-4273-80de-84706acf064a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.774268] env[62558]: DEBUG nova.compute.manager [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing instance network info cache due to event network-changed-24332129-4159-4273-80de-84706acf064a. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.774447] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.846416] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.846668] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.987482] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267372, 'name': ReconfigVM_Task, 'duration_secs': 0.260132} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.987843] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-393854fc-2de5-43ff-8f13-b76ec330d217/volume-393854fc-2de5-43ff-8f13-b76ec330d217.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.988019] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.042876] env[62558]: WARNING nova.network.neutron [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] 63d6d90f-c869-40f3-8247-c3f519646f00 already exists in list: networks containing: ['63d6d90f-c869-40f3-8247-c3f519646f00']. ignoring it [ 1093.300890] env[62558]: DEBUG nova.network.neutron [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24332129-4159-4273-80de-84706acf064a", "address": "fa:16:3e:ba:fc:4a", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24332129-41", "ovs_interfaceid": "24332129-4159-4273-80de-84706acf064a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.351837] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.351969] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1093.496452] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112ea1f1-4bec-4c28-bf78-1a10e70e6781 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.515062] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66c44fb-8271-4e82-ac28-403ac66c0468 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.531657] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.803634] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.804393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.804556] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.804849] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.805043] env[62558]: DEBUG nova.network.neutron [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Refreshing network info cache for port 24332129-4159-4273-80de-84706acf064a {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1093.807161] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66897b79-673a-4df6-aa84-1b6e73c53d44 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.823482] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1093.823704] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1093.823863] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1093.824057] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1093.824212] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1093.824363] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1093.824565] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1093.824725] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1093.824889] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1093.825064] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1093.825242] env[62558]: DEBUG nova.virt.hardware [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.831319] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfiguring VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1093.832136] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-511d8754-0440-46a0-99b4-4e654ad8c8fd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.848910] env[62558]: DEBUG oslo_vmware.api [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1093.848910] env[62558]: value = "task-1267373" [ 1093.848910] env[62558]: _type = "Task" [ 1093.848910] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.857234] env[62558]: DEBUG oslo_vmware.api [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267373, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.359885] env[62558]: DEBUG oslo_vmware.api [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267373, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.504434] env[62558]: DEBUG nova.network.neutron [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updated VIF entry in instance network info cache for port 24332129-4159-4273-80de-84706acf064a. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1094.504913] env[62558]: DEBUG nova.network.neutron [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24332129-4159-4273-80de-84706acf064a", "address": "fa:16:3e:ba:fc:4a", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24332129-41", "ovs_interfaceid": "24332129-4159-4273-80de-84706acf064a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.858933] env[62558]: DEBUG oslo_vmware.api [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267373, 'name': ReconfigVM_Task, 'duration_secs': 0.541543} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.859414] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.859625] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfigured VM to attach interface {{(pid=62558) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1095.008159] env[62558]: DEBUG oslo_concurrency.lockutils [req-fcb3f0e7-bb1f-41ea-94e5-6907033acf2c req-59b21d2a-9663-4677-a465-3d016460b288 service nova] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.173271] env[62558]: DEBUG nova.network.neutron [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Port f7bda9aa-e43a-4d03-a623-e549e9529118 binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1095.364426] env[62558]: DEBUG oslo_concurrency.lockutils [None req-fedb0bff-4a2c-4ab9-8150-69a1a7d38f67 tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 5.956s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.895192] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.895366] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.895511] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1096.198918] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.199139] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.199393] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.595760] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.596079] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.079569] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [{"id": "6fcb24d4-f48e-4628-995d-10f53f01154b", "address": "fa:16:3e:09:d7:9e", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb24d4-f4", "ovs_interfaceid": "6fcb24d4-f48e-4628-995d-10f53f01154b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.098824] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.099018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.099905] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283b344d-a04f-4e3f-84fd-afccc2168444 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.118850] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ffecbb-7a4f-4ed9-9ee0-91fb42228be8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.144319] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfiguring VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1097.144596] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe597991-7154-4e5b-9baa-bda2516fe8bf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.162614] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1097.162614] env[62558]: value = "task-1267374" [ 1097.162614] env[62558]: _type = "Task" [ 1097.162614] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.170238] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.229543] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.229749] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.229929] env[62558]: DEBUG nova.network.neutron [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.582020] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.582292] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1097.582422] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.582592] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.582742] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.582883] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.583055] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.672942] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.064560] env[62558]: DEBUG nova.network.neutron [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.086264] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.086561] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.086760] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.086920] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1098.087869] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb9f5c-e63b-4983-b1e6-d8a6275a83f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.096276] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba0ae99-5e8c-41db-8488-84f001f61f19 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.110203] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7566fa-5e3c-4119-ba41-472e99f04c0a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.116895] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f978376-9fff-4621-8291-77e6524158a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.147050] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180468MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1098.147200] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.147410] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.172999] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.567461] env[62558]: DEBUG oslo_concurrency.lockutils [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.674752] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.696927] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.697162] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.697440] env[62558]: DEBUG nova.compute.manager [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.698161] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a91b9c7-8c49-4f83-8256-97d42bb0e501 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.704691] env[62558]: DEBUG nova.compute.manager [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1098.705259] env[62558]: DEBUG nova.objects.instance [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'flavor' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.076874] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994e0f3a-ed58-4ee5-b14c-e4e420ca0a20 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.083942] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a876cb8f-6341-4dda-a9fa-a77da2f53e59 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.155416] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Applying migration context for instance 5be1007a-6229-4d45-82d0-53d81a28204c as it has an incoming, in-progress migration 9fd47a8b-ae6a-4ecf-b664-2f191a071d7f. Migration status is post-migrating {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1099.156549] env[62558]: INFO nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating resource usage from migration 9fd47a8b-ae6a-4ecf-b664-2f191a071d7f [ 1099.176389] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.178496] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.178641] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5f75bf07-002f-4ed4-930a-fbe6f939cb4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.178761] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5082012b-e8ed-41d6-a036-b64cbb1e363c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.178879] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance ac1ec03d-e04b-42eb-a1f3-c14887241bfa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.178989] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e26d956-f0f4-4ab1-b23c-fd6859875929 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.179122] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 301dde86-496e-45b7-a86e-5bda587908db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.179237] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Migration 9fd47a8b-ae6a-4ecf-b664-2f191a071d7f is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1099.179387] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5be1007a-6229-4d45-82d0-53d81a28204c actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.179520] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1099.179650] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1099.212944] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.213245] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d92d4d8-0642-468e-960f-692ce96b15c9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.221362] env[62558]: DEBUG oslo_vmware.api [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1099.221362] env[62558]: value = "task-1267375" [ 1099.221362] env[62558]: _type = "Task" [ 1099.221362] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.232146] env[62558]: DEBUG oslo_vmware.api [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.288410] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbeeaef-bfc5-4f02-948b-9c34fbf066c3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.295629] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5570f3d6-35e4-4a74-a987-7146e6b2b1bc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.325243] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda21084-5577-4e7f-b8b0-d1994b89b9b8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.333379] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85870539-5295-4f89-a100-9c50a090f413 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.347021] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.675882] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.730338] env[62558]: DEBUG oslo_vmware.api [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267375, 'name': PowerOffVM_Task, 'duration_secs': 0.157061} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.730590] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.730769] env[62558]: DEBUG nova.compute.manager [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.731514] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df3554a-8828-4b9b-bc91-ee607077b00d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.849874] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.175778] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.177240] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6956facd-633c-40bd-9566-e56be29d038d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.194178] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5934ec-9a46-4307-b531-d22d6832c412 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.200233] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1100.242934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3c821ded-1edd-44c4-9ed6-ff6cb12a3467 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.354800] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1100.354974] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.208s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.677020] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.706633] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.706937] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51178e93-a663-445d-a516-e41746ee5c71 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.713954] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1100.713954] env[62558]: value = "task-1267376" [ 1100.713954] env[62558]: _type = "Task" [ 1100.713954] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.721370] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267376, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.057433] env[62558]: DEBUG nova.objects.instance [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'flavor' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.178088] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.223380] env[62558]: DEBUG oslo_vmware.api [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267376, 'name': PowerOnVM_Task, 'duration_secs': 0.350265} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.223650] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.223838] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf1d4ca-03db-4f31-9bb9-ee4479df8988 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance '5be1007a-6229-4d45-82d0-53d81a28204c' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1101.562227] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.562409] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.562574] env[62558]: DEBUG nova.network.neutron [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1101.562756] env[62558]: DEBUG nova.objects.instance [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'info_cache' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.677939] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.065923] env[62558]: DEBUG nova.objects.base [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Object Instance<4e26d956-f0f4-4ab1-b23c-fd6859875929> lazy-loaded attributes: flavor,info_cache {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1102.178650] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.678384] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.765587] env[62558]: DEBUG nova.network.neutron [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.184463] env[62558]: DEBUG oslo_vmware.api [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267374, 'name': ReconfigVM_Task, 'duration_secs': 5.742497} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.184851] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.185226] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Reconfigured VM to detach interface {{(pid=62558) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1103.275593] env[62558]: DEBUG oslo_concurrency.lockutils [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.779223] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.779693] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-022c2462-197e-4833-a16f-8f77eeff12cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.786937] env[62558]: DEBUG oslo_vmware.api [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1103.786937] env[62558]: value = "task-1267377" [ 1103.786937] env[62558]: _type = "Task" [ 1103.786937] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.794622] env[62558]: DEBUG oslo_vmware.api [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267377, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.092720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.092962] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.093258] env[62558]: DEBUG nova.compute.manager [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Going to confirm migration 4 {{(pid=62558) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1104.296819] env[62558]: DEBUG oslo_vmware.api [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267377, 'name': PowerOnVM_Task, 'duration_secs': 0.378906} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.297101] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.297290] env[62558]: DEBUG nova.compute.manager [None req-f89ca801-5d35-4b8d-a103-d07f12b04aa7 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.298064] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c07680-7aeb-43cb-bfbf-e13335ff3100 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.511309] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.511496] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquired lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.511665] env[62558]: DEBUG nova.network.neutron [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.645978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.646177] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquired lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.646399] env[62558]: DEBUG nova.network.neutron [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.646611] env[62558]: DEBUG nova.objects.instance [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'info_cache' on Instance uuid 5be1007a-6229-4d45-82d0-53d81a28204c {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.001377] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.001801] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.001844] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.002170] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.002279] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.004440] env[62558]: INFO nova.compute.manager [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Terminating instance [ 1105.006271] env[62558]: DEBUG nova.compute.manager [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1105.006486] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.007321] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466735dc-de17-44ee-8d8e-099f3c07d411 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.016782] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.017325] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a758090-92f1-4cbc-9944-2f2932bd560a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.023472] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1105.023472] env[62558]: value = "task-1267378" [ 1105.023472] env[62558]: _type = "Task" [ 1105.023472] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.031348] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267378, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.310110] env[62558]: INFO nova.network.neutron [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Port 24332129-4159-4273-80de-84706acf064a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1105.310493] env[62558]: DEBUG nova.network.neutron [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [{"id": "66f313e4-07aa-4e57-b472-a4408da24e26", "address": "fa:16:3e:b6:5d:0f", "network": {"id": "63d6d90f-c869-40f3-8247-c3f519646f00", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2069680511-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd56f6c392f14aa880ef4891990c1d44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "715e3f37-7401-48fb-a0ee-59d340b40de1", "external-id": "nsx-vlan-transportzone-739", "segmentation_id": 739, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66f313e4-07", "ovs_interfaceid": "66f313e4-07aa-4e57-b472-a4408da24e26", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.533600] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267378, 'name': PowerOffVM_Task, 'duration_secs': 0.194379} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.533760] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.533944] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.534184] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25cb16d4-4156-42e2-9e03-723d2b68a9d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.591894] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1105.592131] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1105.592314] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleting the datastore file [datastore2] ac1ec03d-e04b-42eb-a1f3-c14887241bfa {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.592575] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eccee7ea-71fa-4477-b14a-10bf961ebdb8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.599087] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1105.599087] env[62558]: value = "task-1267380" [ 1105.599087] env[62558]: _type = "Task" [ 1105.599087] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.606588] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267380, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.813745] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Releasing lock "refresh_cache-ac1ec03d-e04b-42eb-a1f3-c14887241bfa" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.109990] env[62558]: DEBUG oslo_vmware.api [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267380, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144749} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.110458] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.110522] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1106.111018] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1106.111018] env[62558]: INFO nova.compute.manager [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1106.111522] env[62558]: DEBUG oslo.service.loopingcall [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.111522] env[62558]: DEBUG nova.compute.manager [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1106.111522] env[62558]: DEBUG nova.network.neutron [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1106.213385] env[62558]: DEBUG nova.network.neutron [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [{"id": "f7bda9aa-e43a-4d03-a623-e549e9529118", "address": "fa:16:3e:d2:bb:af", "network": {"id": "3f901ddf-bce0-4fbf-8b9a-c00e97b7235e", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1473495815-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78b69738b90142a4943069006e349445", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ff81f9-72b2-4e58-a8d8-5699907f7459", "external-id": "nsx-vlan-transportzone-978", "segmentation_id": 978, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7bda9aa-e4", "ovs_interfaceid": "f7bda9aa-e43a-4d03-a623-e549e9529118", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.317437] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0e44f519-84ae-4465-8fe6-4d14147d714e tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "interface-ac1ec03d-e04b-42eb-a1f3-c14887241bfa-24332129-4159-4273-80de-84706acf064a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.721s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.716611] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Releasing lock "refresh_cache-5be1007a-6229-4d45-82d0-53d81a28204c" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.716923] env[62558]: DEBUG nova.objects.instance [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'migration_context' on Instance uuid 5be1007a-6229-4d45-82d0-53d81a28204c {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.755976] env[62558]: DEBUG nova.compute.manager [req-9cfbd86a-6902-43de-9230-0eeb804a7a69 req-571491f4-dbf8-4a46-8b11-0071f37eccbc service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Received event network-vif-deleted-66f313e4-07aa-4e57-b472-a4408da24e26 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.756240] env[62558]: INFO nova.compute.manager [req-9cfbd86a-6902-43de-9230-0eeb804a7a69 req-571491f4-dbf8-4a46-8b11-0071f37eccbc service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Neutron deleted interface 66f313e4-07aa-4e57-b472-a4408da24e26; detaching it from the instance and deleting it from the info cache [ 1106.756448] env[62558]: DEBUG nova.network.neutron [req-9cfbd86a-6902-43de-9230-0eeb804a7a69 req-571491f4-dbf8-4a46-8b11-0071f37eccbc service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.220172] env[62558]: DEBUG nova.objects.base [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Object Instance<5be1007a-6229-4d45-82d0-53d81a28204c> lazy-loaded attributes: info_cache,migration_context {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1107.220814] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be870435-0586-4b39-8dc5-5f8aee4ba91e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.239244] env[62558]: DEBUG nova.network.neutron [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.243041] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b529bdd5-774d-42a9-9a39-5588711c695e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.248091] env[62558]: DEBUG oslo_vmware.api [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1107.248091] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52468ca7-79b3-a392-0a77-3e9ca6f102fb" [ 1107.248091] env[62558]: _type = "Task" [ 1107.248091] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.257292] env[62558]: DEBUG oslo_vmware.api [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52468ca7-79b3-a392-0a77-3e9ca6f102fb, 'name': SearchDatastore_Task, 'duration_secs': 0.006905} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.257610] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.257973] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.260430] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a781f40-e1f8-4639-9b44-610620142dcf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.269852] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92126c09-28fb-4ddd-be61-459a112182e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.287216] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.287461] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.299531] env[62558]: DEBUG nova.compute.manager [req-9cfbd86a-6902-43de-9230-0eeb804a7a69 req-571491f4-dbf8-4a46-8b11-0071f37eccbc service nova] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Detach interface failed, port_id=66f313e4-07aa-4e57-b472-a4408da24e26, reason: Instance ac1ec03d-e04b-42eb-a1f3-c14887241bfa could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1107.745831] env[62558]: INFO nova.compute.manager [-] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Took 1.63 seconds to deallocate network for instance. [ 1107.790286] env[62558]: DEBUG nova.compute.utils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.859757] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66912b7a-bf29-4741-a566-423691b1719f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.867730] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca059eae-bbd0-4567-818f-fa0f8d89e155 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.896647] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bdfdc9-efbf-4212-8e52-e51171ddd9e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.903894] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b62f82a-ab40-4bf9-adaf-9be9d0d1a273 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.916369] env[62558]: DEBUG nova.compute.provider_tree [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.253136] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.293231] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.419793] env[62558]: DEBUG nova.scheduler.client.report [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.353378] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.353729] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.353935] env[62558]: INFO nova.compute.manager [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Attaching volume 982dafca-f8b5-40c5-a8dd-668cab0b525e to /dev/sdb [ 1109.388048] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600adccd-0fc6-4721-a69e-94884d4ee9e2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.395105] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb8aeb6-029b-44b8-9804-4155835440ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.408401] env[62558]: DEBUG nova.virt.block_device [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating existing volume attachment record: 19684fd9-e2a1-402e-85b9-94f8a5e4cdd3 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1109.429717] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.172s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.432442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.180s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.432683] env[62558]: DEBUG nova.objects.instance [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'resources' on Instance uuid ac1ec03d-e04b-42eb-a1f3-c14887241bfa {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.998086] env[62558]: INFO nova.scheduler.client.report [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocation for migration 9fd47a8b-ae6a-4ecf-b664-2f191a071d7f [ 1110.052843] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc43652-ca10-43a0-9d2f-7587059d62ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.061045] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe799e3f-1e80-41ac-b5b5-55571696818d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.091127] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6842ef-124a-4651-9981-bdcb77d9ccff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.099539] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34829881-a5fd-4268-9444-2404b77345db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.114662] env[62558]: DEBUG nova.compute.provider_tree [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.506083] env[62558]: DEBUG oslo_concurrency.lockutils [None req-70a922b8-50ac-4468-bb0f-a6adfc5a0030 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.413s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.618098] env[62558]: DEBUG nova.scheduler.client.report [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.758889] env[62558]: INFO nova.compute.manager [None req-25238a01-3725-4191-adae-d01c78ba8d14 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Get console output [ 1110.759224] env[62558]: WARNING nova.virt.vmwareapi.driver [None req-25238a01-3725-4191-adae-d01c78ba8d14 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] The console log is missing. Check your VSPC configuration [ 1111.122715] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.142199] env[62558]: INFO nova.scheduler.client.report [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted allocations for instance ac1ec03d-e04b-42eb-a1f3-c14887241bfa [ 1111.649669] env[62558]: DEBUG oslo_concurrency.lockutils [None req-bb20ebc2-dc8a-4d64-bfee-b1b14ff9d9be tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "ac1ec03d-e04b-42eb-a1f3-c14887241bfa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.648s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.879047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.879047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.879047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.879047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.879047] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.881118] env[62558]: INFO nova.compute.manager [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Terminating instance [ 1111.882756] env[62558]: DEBUG nova.compute.manager [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1111.882984] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1111.883831] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17757616-e645-4e5c-94cc-aaf5a9e08f52 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.890919] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.891161] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e525e6ec-4a36-4c59-91b6-3dd81751aeb7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.896190] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1111.896190] env[62558]: value = "task-1267383" [ 1111.896190] env[62558]: _type = "Task" [ 1111.896190] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.903457] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.406091] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267383, 'name': PowerOffVM_Task, 'duration_secs': 0.149059} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.406754] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.406933] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1112.407210] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0a57faf-6baf-43c3-a313-497c4ed95585 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.466242] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1112.466590] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1112.466697] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleting the datastore file [datastore1] 5f75bf07-002f-4ed4-930a-fbe6f939cb4f {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1112.466978] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-815578c5-89f3-47a1-933f-099c46d3d426 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.472889] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for the task: (returnval){ [ 1112.472889] env[62558]: value = "task-1267385" [ 1112.472889] env[62558]: _type = "Task" [ 1112.472889] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.480312] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.982388] env[62558]: DEBUG oslo_vmware.api [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Task: {'id': task-1267385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09089} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.982651] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.982950] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.983164] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.983343] env[62558]: INFO nova.compute.manager [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1112.983647] env[62558]: DEBUG oslo.service.loopingcall [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1112.983768] env[62558]: DEBUG nova.compute.manager [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1112.983863] env[62558]: DEBUG nova.network.neutron [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1113.246708] env[62558]: DEBUG nova.compute.manager [req-030efb98-17b4-4e71-bb0e-1b6b22e9997d req-9dec3891-99f2-4bc2-bfeb-253106728763 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Received event network-vif-deleted-dedfab3d-cdf6-40b7-953a-7d13dad8f5a6 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.246853] env[62558]: INFO nova.compute.manager [req-030efb98-17b4-4e71-bb0e-1b6b22e9997d req-9dec3891-99f2-4bc2-bfeb-253106728763 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Neutron deleted interface dedfab3d-cdf6-40b7-953a-7d13dad8f5a6; detaching it from the instance and deleting it from the info cache [ 1113.247010] env[62558]: DEBUG nova.network.neutron [req-030efb98-17b4-4e71-bb0e-1b6b22e9997d req-9dec3891-99f2-4bc2-bfeb-253106728763 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.726150] env[62558]: DEBUG nova.network.neutron [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.749136] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dd3228d-72ce-42bf-8411-7bb93e360ff3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.758417] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb76b01-6eaa-49f9-b9c3-80ffa6a3dc3d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.783311] env[62558]: DEBUG nova.compute.manager [req-030efb98-17b4-4e71-bb0e-1b6b22e9997d req-9dec3891-99f2-4bc2-bfeb-253106728763 service nova] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Detach interface failed, port_id=dedfab3d-cdf6-40b7-953a-7d13dad8f5a6, reason: Instance 5f75bf07-002f-4ed4-930a-fbe6f939cb4f could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1113.952261] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1113.952512] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272626', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'name': 'volume-982dafca-f8b5-40c5-a8dd-668cab0b525e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '301dde86-496e-45b7-a86e-5bda587908db', 'attached_at': '', 'detached_at': '', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'serial': '982dafca-f8b5-40c5-a8dd-668cab0b525e'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1113.953447] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c09558-03a6-44e2-b58d-1610abd726ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.969420] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d83b2e-52a8-4c12-b92e-e9c06bcf9918 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.992481] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-982dafca-f8b5-40c5-a8dd-668cab0b525e/volume-982dafca-f8b5-40c5-a8dd-668cab0b525e.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.992795] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bca3a9ee-0b65-4170-96be-1e26e4fc2ea6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.012360] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1114.012360] env[62558]: value = "task-1267386" [ 1114.012360] env[62558]: _type = "Task" [ 1114.012360] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.019949] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267386, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.229078] env[62558]: INFO nova.compute.manager [-] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Took 1.24 seconds to deallocate network for instance. [ 1114.521612] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267386, 'name': ReconfigVM_Task, 'duration_secs': 0.386032} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.521900] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-982dafca-f8b5-40c5-a8dd-668cab0b525e/volume-982dafca-f8b5-40c5-a8dd-668cab0b525e.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.526471] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c068e5d-3903-4490-ad25-1c5e945e57c0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.541496] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1114.541496] env[62558]: value = "task-1267387" [ 1114.541496] env[62558]: _type = "Task" [ 1114.541496] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.549120] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267387, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.735767] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.736065] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.736297] env[62558]: DEBUG nova.objects.instance [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lazy-loading 'resources' on Instance uuid 5f75bf07-002f-4ed4-930a-fbe6f939cb4f {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.051644] env[62558]: DEBUG oslo_vmware.api [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267387, 'name': ReconfigVM_Task, 'duration_secs': 0.139491} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.051941] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272626', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'name': 'volume-982dafca-f8b5-40c5-a8dd-668cab0b525e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '301dde86-496e-45b7-a86e-5bda587908db', 'attached_at': '', 'detached_at': '', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'serial': '982dafca-f8b5-40c5-a8dd-668cab0b525e'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1115.315258] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afd56e1-dcd9-4327-bead-960dab229550 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.322625] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73ec170-e602-4674-8112-8fd01e71a7e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.351727] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abba416a-c4b8-49a6-adf1-72a86c549811 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.358253] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a740914-9bf0-47ef-933b-e2654c7b148f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.370607] env[62558]: DEBUG nova.compute.provider_tree [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.873159] env[62558]: DEBUG nova.scheduler.client.report [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1116.152360] env[62558]: DEBUG nova.objects.instance [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid 301dde86-496e-45b7-a86e-5bda587908db {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.377447] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.400852] env[62558]: INFO nova.scheduler.client.report [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Deleted allocations for instance 5f75bf07-002f-4ed4-930a-fbe6f939cb4f [ 1116.656988] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3104f298-b6c5-4df3-8da1-af4801e7491c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.303s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.799881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.800176] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.907963] env[62558]: DEBUG oslo_concurrency.lockutils [None req-561a5f0c-8905-4a9f-85b7-37b167405a4c tempest-AttachInterfacesTestJSON-761087401 tempest-AttachInterfacesTestJSON-761087401-project-member] Lock "5f75bf07-002f-4ed4-930a-fbe6f939cb4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.029s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.303251] env[62558]: INFO nova.compute.manager [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Detaching volume 982dafca-f8b5-40c5-a8dd-668cab0b525e [ 1117.341185] env[62558]: INFO nova.virt.block_device [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Attempting to driver detach volume 982dafca-f8b5-40c5-a8dd-668cab0b525e from mountpoint /dev/sdb [ 1117.341458] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1117.342278] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272626', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'name': 'volume-982dafca-f8b5-40c5-a8dd-668cab0b525e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '301dde86-496e-45b7-a86e-5bda587908db', 'attached_at': '', 'detached_at': '', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'serial': '982dafca-f8b5-40c5-a8dd-668cab0b525e'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1117.342544] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dfada9-1bff-4da6-a8c7-3f1eb54e0bf2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.364335] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f2595b-e50d-4621-8062-3ea5acf029be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.371109] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418950c5-e20a-4011-ab87-ea13c4c84961 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.390569] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e88dcf-284e-4848-a76c-ef7f88d704cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.406905] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] The volume has not been displaced from its original location: [datastore1] volume-982dafca-f8b5-40c5-a8dd-668cab0b525e/volume-982dafca-f8b5-40c5-a8dd-668cab0b525e.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1117.412045] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.412559] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f22752f5-1505-40d6-96a9-995803112dd2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.440603] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1117.440603] env[62558]: value = "task-1267388" [ 1117.440603] env[62558]: _type = "Task" [ 1117.440603] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.451344] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267388, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.951824] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267388, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.451313] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267388, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.952220] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267388, 'name': ReconfigVM_Task, 'duration_secs': 1.210616} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.952520] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1118.957129] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e5437cd-3f6d-470e-b72f-c0ef3e326c66 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.972466] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1118.972466] env[62558]: value = "task-1267389" [ 1118.972466] env[62558]: _type = "Task" [ 1118.972466] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.980068] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267389, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.482649] env[62558]: DEBUG oslo_vmware.api [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267389, 'name': ReconfigVM_Task, 'duration_secs': 0.139501} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.482954] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272626', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'name': 'volume-982dafca-f8b5-40c5-a8dd-668cab0b525e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '301dde86-496e-45b7-a86e-5bda587908db', 'attached_at': '', 'detached_at': '', 'volume_id': '982dafca-f8b5-40c5-a8dd-668cab0b525e', 'serial': '982dafca-f8b5-40c5-a8dd-668cab0b525e'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1119.998122] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.026056] env[62558]: DEBUG nova.objects.instance [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid 301dde86-496e-45b7-a86e-5bda587908db {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.501240] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Getting list of instances from cluster (obj){ [ 1120.501240] env[62558]: value = "domain-c8" [ 1120.501240] env[62558]: _type = "ClusterComputeResource" [ 1120.501240] env[62558]: } {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1120.502354] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcf3516-daa1-47cb-b09f-7480a8bcd32f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.516670] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Got total of 5 instances {{(pid=62558) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1120.516875] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.517086] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 5082012b-e8ed-41d6-a036-b64cbb1e363c {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.517249] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.517400] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 301dde86-496e-45b7-a86e-5bda587908db {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.517547] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Triggering sync for uuid 5be1007a-6229-4d45-82d0-53d81a28204c {{(pid=62558) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1120.517906] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.518146] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.518412] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.518600] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.518836] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.519033] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.519276] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.519485] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.519661] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.520463] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3494bdcb-aa16-46fc-b126-0fab28446123 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.523230] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d122d3-ea67-48af-8c12-b409926dbcfc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.525967] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1056b6-9ec9-42fa-b835-70917fa7f61a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.528693] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d25ada-d3e8-4dd9-86ec-c8f42edab398 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.040509] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b5bbdd96-62d1-4b7f-b112-c5df710de053 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.240s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.041713] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.522s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.044426] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c71394-94ba-4e41-ac20-760c2182099b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.047293] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.529s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.047603] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.529s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.047889] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.530s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.049742] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.530s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.554081] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.083113] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.083416] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.083636] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "301dde86-496e-45b7-a86e-5bda587908db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.083834] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.084074] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.086154] env[62558]: INFO nova.compute.manager [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Terminating instance [ 1122.087923] env[62558]: DEBUG nova.compute.manager [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1122.088149] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1122.088957] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885cbf51-5fdc-448b-9e3e-4d291f670b66 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.096413] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.096925] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7b0d4fa-f8f5-4c1f-ad9b-34f204b0d455 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.102630] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1122.102630] env[62558]: value = "task-1267390" [ 1122.102630] env[62558]: _type = "Task" [ 1122.102630] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.110487] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.612552] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267390, 'name': PowerOffVM_Task, 'duration_secs': 0.182399} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.612826] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1122.613008] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1122.613276] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d7d0c40-54b8-42dd-bf4e-1dc78efeb65a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.668951] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1122.669164] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1122.669352] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleting the datastore file [datastore1] 301dde86-496e-45b7-a86e-5bda587908db {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1122.669608] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aec50d76-18ff-498c-9283-3cdcf5e6bce6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.676235] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1122.676235] env[62558]: value = "task-1267392" [ 1122.676235] env[62558]: _type = "Task" [ 1122.676235] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.683828] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.185483] env[62558]: DEBUG oslo_vmware.api [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134251} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.185732] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.185918] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1123.186112] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1123.186290] env[62558]: INFO nova.compute.manager [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1123.186532] env[62558]: DEBUG oslo.service.loopingcall [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.186734] env[62558]: DEBUG nova.compute.manager [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1123.186843] env[62558]: DEBUG nova.network.neutron [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1123.738693] env[62558]: DEBUG nova.compute.manager [req-e5a7dc4d-e0ec-461e-ad00-792975eb1b1b req-ff8dddeb-d993-45a4-a96c-550fd321ebfb service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Received event network-vif-deleted-7c11b24d-6b71-48d9-ae68-a7c4f88c7367 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.738693] env[62558]: INFO nova.compute.manager [req-e5a7dc4d-e0ec-461e-ad00-792975eb1b1b req-ff8dddeb-d993-45a4-a96c-550fd321ebfb service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Neutron deleted interface 7c11b24d-6b71-48d9-ae68-a7c4f88c7367; detaching it from the instance and deleting it from the info cache [ 1123.738693] env[62558]: DEBUG nova.network.neutron [req-e5a7dc4d-e0ec-461e-ad00-792975eb1b1b req-ff8dddeb-d993-45a4-a96c-550fd321ebfb service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.098474] env[62558]: DEBUG nova.network.neutron [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.240542] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b18b52a-ba1e-462a-b586-7bfa7184a48f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.249648] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bb3171-ce0d-4e8f-bf88-03df26e0c1c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.275341] env[62558]: DEBUG nova.compute.manager [req-e5a7dc4d-e0ec-461e-ad00-792975eb1b1b req-ff8dddeb-d993-45a4-a96c-550fd321ebfb service nova] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Detach interface failed, port_id=7c11b24d-6b71-48d9-ae68-a7c4f88c7367, reason: Instance 301dde86-496e-45b7-a86e-5bda587908db could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1124.601130] env[62558]: INFO nova.compute.manager [-] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Took 1.41 seconds to deallocate network for instance. [ 1124.630771] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.630957] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1125.108081] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.108521] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.108608] env[62558]: DEBUG nova.objects.instance [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'resources' on Instance uuid 301dde86-496e-45b7-a86e-5bda587908db {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.630927] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.631098] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1125.690025] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918452e6-aace-41dd-b197-0b6b9fb6a581 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.696787] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470b85d5-ff65-4310-85fd-27f2a77d4183 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.726158] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a795be3-ef86-4093-bd26-892ac9e1b2e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.732720] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d62fc45-bd5a-4275-af1c-1bc7b61c0d6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.745194] env[62558]: DEBUG nova.compute.provider_tree [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.247786] env[62558]: DEBUG nova.scheduler.client.report [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.151552] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Didn't find any instances for network info cache update. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1129.151842] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.151959] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.152133] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.152296] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.254400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.146s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.280663] env[62558]: INFO nova.scheduler.client.report [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted allocations for instance 301dde86-496e-45b7-a86e-5bda587908db [ 1129.656352] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.656585] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.656747] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.656957] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1129.658037] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af0bf2f-2ee0-4cc5-b95f-78d27cf04c17 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.666255] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ebe1ea-fe02-405b-86af-aae1afc76c5e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.679742] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6e2691-e514-4041-8567-ab1f7ac4f59a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.685669] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428403dd-97c0-4dec-80b7-e66a9117b6d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.714851] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180468MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1129.715014] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.715226] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.788868] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a6cb255b-bf4c-434d-a8dd-6d60a7809c2c tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "301dde86-496e-45b7-a86e-5bda587908db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.705s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.745880] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.746188] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5082012b-e8ed-41d6-a036-b64cbb1e363c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.746188] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e26d956-f0f4-4ab1-b23c-fd6859875929 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.746449] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 5be1007a-6229-4d45-82d0-53d81a28204c actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1130.746449] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1130.746592] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1130.810964] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cc01d9-af92-433c-8c58-8b1eba14e219 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.818512] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d6940e-c515-4258-96ce-302d155495cc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.847372] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a11a69-2868-4596-93a8-950104add610 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.854164] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72875209-5e9a-4044-beb1-f0424614b038 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.867449] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.370398] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.875526] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1131.875987] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.160s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.137592] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.137827] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.354210] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.354420] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.354579] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.640309] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1133.164342] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.164619] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.166211] env[62558]: INFO nova.compute.claims [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.248155] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97371d0-2f7d-4e08-a114-abcb981b47f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.255405] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae01c1cc-bb00-4cfb-917a-97b0afadcfe0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.285646] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db43905-1e71-4470-85ea-5e8a02837b4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.292214] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9076410-6c62-437b-afb8-7f12f62b0fc1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.304854] env[62558]: DEBUG nova.compute.provider_tree [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.808444] env[62558]: DEBUG nova.scheduler.client.report [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.313558] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.314110] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1135.819214] env[62558]: DEBUG nova.compute.utils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1135.820715] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1135.820886] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1135.867504] env[62558]: DEBUG nova.policy [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37bb028d611e42d98319c4afba5b1b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '46aa5789fccf4925aa49f9c4050c8463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1136.249044] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Successfully created port: eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1136.324159] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1137.335062] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1137.383119] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1137.383409] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1137.383569] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1137.383753] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1137.383902] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1137.384063] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1137.384285] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1137.384488] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1137.384670] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1137.384834] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1137.385011] env[62558]: DEBUG nova.virt.hardware [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.385857] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549b4e5f-411a-4416-b18d-29fbf0168161 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.393461] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f445154f-36a6-4748-8a14-7390259c655a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.685078] env[62558]: DEBUG nova.compute.manager [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Received event network-vif-plugged-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.685379] env[62558]: DEBUG oslo_concurrency.lockutils [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.685602] env[62558]: DEBUG oslo_concurrency.lockutils [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.685792] env[62558]: DEBUG oslo_concurrency.lockutils [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.685942] env[62558]: DEBUG nova.compute.manager [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] No waiting events found dispatching network-vif-plugged-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1137.686193] env[62558]: WARNING nova.compute.manager [req-91161c34-c2d1-4a2a-9e1a-6d69aa88eb6f req-8cf8b051-15e9-4332-9520-d92b19c59b6b service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Received unexpected event network-vif-plugged-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c for instance with vm_state building and task_state spawning. [ 1137.689135] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.689346] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.689542] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.689720] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.689883] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.691738] env[62558]: INFO nova.compute.manager [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Terminating instance [ 1137.693494] env[62558]: DEBUG nova.compute.manager [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.693704] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.694174] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f25371df-95da-4d86-8fa8-fafa32c17f53 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.700994] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1137.700994] env[62558]: value = "task-1267393" [ 1137.700994] env[62558]: _type = "Task" [ 1137.700994] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.709066] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.770283] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Successfully updated port: eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1138.211030] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.272018] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.272234] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.272383] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.711030] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267393, 'name': PowerOffVM_Task, 'duration_secs': 0.910962} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.711445] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.711528] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1138.711684] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272623', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'name': 'volume-393854fc-2de5-43ff-8f13-b76ec330d217', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5be1007a-6229-4d45-82d0-53d81a28204c', 'attached_at': '2024-10-11T05:09:50.000000', 'detached_at': '', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'serial': '393854fc-2de5-43ff-8f13-b76ec330d217'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1138.712445] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423dd410-92e5-4929-89f7-25a060311d4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.729579] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d8d946-524f-42c1-a6c2-8b328e1baf89 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.735289] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7344406-d0e3-482b-bb74-017803c49d80 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.751886] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7489d0e-eff2-46ce-bb60-56783425b1d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.768359] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] The volume has not been displaced from its original location: [datastore1] volume-393854fc-2de5-43ff-8f13-b76ec330d217/volume-393854fc-2de5-43ff-8f13-b76ec330d217.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1138.773509] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1138.773728] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd8a2124-f303-4354-87f0-ee509961db9e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.792018] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1138.792018] env[62558]: value = "task-1267394" [ 1138.792018] env[62558]: _type = "Task" [ 1138.792018] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.799359] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267394, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.813767] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1138.940525] env[62558]: DEBUG nova.network.neutron [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating instance_info_cache with network_info: [{"id": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "address": "fa:16:3e:2f:d9:bf", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaa9fa2a-20", "ovs_interfaceid": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.302135] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267394, 'name': ReconfigVM_Task, 'duration_secs': 0.203552} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.302447] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1139.307136] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de691789-ebb1-4e4b-8a73-c19913a86df4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.321500] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1139.321500] env[62558]: value = "task-1267395" [ 1139.321500] env[62558]: _type = "Task" [ 1139.321500] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.332411] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267395, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.442798] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.443173] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Instance network_info: |[{"id": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "address": "fa:16:3e:2f:d9:bf", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaa9fa2a-20", "ovs_interfaceid": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1139.443587] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:d9:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359c2c31-99c4-41d7-a513-3bc4825897a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.450797] env[62558]: DEBUG oslo.service.loopingcall [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.451010] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.451240] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2cc830c-d6aa-4852-a26b-445f444ea4f0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.470205] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.470205] env[62558]: value = "task-1267396" [ 1139.470205] env[62558]: _type = "Task" [ 1139.470205] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.477512] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267396, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.711189] env[62558]: DEBUG nova.compute.manager [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Received event network-changed-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.711503] env[62558]: DEBUG nova.compute.manager [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Refreshing instance network info cache due to event network-changed-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1139.711652] env[62558]: DEBUG oslo_concurrency.lockutils [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] Acquiring lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.711879] env[62558]: DEBUG oslo_concurrency.lockutils [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] Acquired lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.712074] env[62558]: DEBUG nova.network.neutron [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Refreshing network info cache for port eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1139.830233] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267395, 'name': ReconfigVM_Task, 'duration_secs': 0.105119} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.830538] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272623', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'name': 'volume-393854fc-2de5-43ff-8f13-b76ec330d217', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '5be1007a-6229-4d45-82d0-53d81a28204c', 'attached_at': '2024-10-11T05:09:50.000000', 'detached_at': '', 'volume_id': '393854fc-2de5-43ff-8f13-b76ec330d217', 'serial': '393854fc-2de5-43ff-8f13-b76ec330d217'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1139.830801] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1139.831546] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093bd639-26cf-47d3-83fe-23d2a078aec7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.837605] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1139.837817] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fb61407-596a-436c-9c50-758d1d2039c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.892799] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1139.893024] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1139.893217] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore1] 5be1007a-6229-4d45-82d0-53d81a28204c {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1139.893466] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94e61f74-de02-4dfd-b471-f2feb5c9f1ec {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.899984] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1139.899984] env[62558]: value = "task-1267398" [ 1139.899984] env[62558]: _type = "Task" [ 1139.899984] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.906851] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.981012] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267396, 'name': CreateVM_Task, 'duration_secs': 0.299536} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.981240] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1139.981900] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.982076] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.982442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1139.982698] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c570e6be-949b-416b-bc49-bce24ff3abb4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.988691] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1139.988691] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527f7d99-f70c-f48c-b185-ab813f122baf" [ 1139.988691] env[62558]: _type = "Task" [ 1139.988691] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.001529] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527f7d99-f70c-f48c-b185-ab813f122baf, 'name': SearchDatastore_Task, 'duration_secs': 0.010673} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.002256] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.002485] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.002717] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.002866] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.003258] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.003676] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c948301-5fa3-421e-8a60-2e59837f4029 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.010759] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.010935] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1140.011825] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96463ccb-dada-4f8a-acc1-8f71149d5167 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.016325] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1140.016325] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521137a8-a8b0-cba4-b5ad-80151e74c6b6" [ 1140.016325] env[62558]: _type = "Task" [ 1140.016325] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.024777] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521137a8-a8b0-cba4-b5ad-80151e74c6b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.386207] env[62558]: DEBUG nova.network.neutron [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updated VIF entry in instance network info cache for port eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1140.386684] env[62558]: DEBUG nova.network.neutron [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating instance_info_cache with network_info: [{"id": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "address": "fa:16:3e:2f:d9:bf", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaa9fa2a-20", "ovs_interfaceid": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.409082] env[62558]: DEBUG oslo_vmware.api [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082219} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.409317] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1140.409499] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1140.409676] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1140.409846] env[62558]: INFO nova.compute.manager [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Took 2.72 seconds to destroy the instance on the hypervisor. [ 1140.410096] env[62558]: DEBUG oslo.service.loopingcall [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.410318] env[62558]: DEBUG nova.compute.manager [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1140.410417] env[62558]: DEBUG nova.network.neutron [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1140.495980] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.496216] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.496449] env[62558]: INFO nova.compute.manager [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Rebooting instance [ 1140.526045] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]521137a8-a8b0-cba4-b5ad-80151e74c6b6, 'name': SearchDatastore_Task, 'duration_secs': 0.012708} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.526703] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9459a2c7-5cb5-477a-b6a0-fed8d9ae0f21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.532624] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1140.532624] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c2c201-ed88-2526-b1ab-356dfdbc5b1a" [ 1140.532624] env[62558]: _type = "Task" [ 1140.532624] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.539985] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c2c201-ed88-2526-b1ab-356dfdbc5b1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.889649] env[62558]: DEBUG oslo_concurrency.lockutils [req-4c8db4c6-8205-452e-add1-1b21c7389a52 req-b6a74335-dadf-475e-bed5-a0bc34727857 service nova] Releasing lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.013122] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.013316] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.013492] env[62558]: DEBUG nova.network.neutron [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1141.043105] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52c2c201-ed88-2526-b1ab-356dfdbc5b1a, 'name': SearchDatastore_Task, 'duration_secs': 0.010527} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.043311] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.043583] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] f6d46374-7a19-4079-8257-1ac69ac82b81/f6d46374-7a19-4079-8257-1ac69ac82b81.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1141.043843] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e3210cf-85d3-4fc2-92bc-049950d0f2ee {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.049852] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1141.049852] env[62558]: value = "task-1267399" [ 1141.049852] env[62558]: _type = "Task" [ 1141.049852] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.057237] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.475677] env[62558]: DEBUG nova.network.neutron [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.559456] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267399, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462453} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.559727] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] f6d46374-7a19-4079-8257-1ac69ac82b81/f6d46374-7a19-4079-8257-1ac69ac82b81.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1141.559941] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1141.560217] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8d60ab5-09bb-4654-ad8b-4cab46e69edd {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.566122] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1141.566122] env[62558]: value = "task-1267400" [ 1141.566122] env[62558]: _type = "Task" [ 1141.566122] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.574253] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267400, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.722237] env[62558]: DEBUG nova.network.neutron [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.737673] env[62558]: DEBUG nova.compute.manager [req-8c31165f-823f-4b9a-ba5e-b19ec166e502 req-05fa3115-f40b-4aac-b3ba-a283dfcbb250 service nova] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Received event network-vif-deleted-f7bda9aa-e43a-4d03-a623-e549e9529118 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1141.979146] env[62558]: INFO nova.compute.manager [-] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Took 1.57 seconds to deallocate network for instance. [ 1142.076430] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267400, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069132} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.076644] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1142.077458] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1309a269-9936-4b95-817f-486d72bba9eb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.099257] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] f6d46374-7a19-4079-8257-1ac69ac82b81/f6d46374-7a19-4079-8257-1ac69ac82b81.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1142.099532] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f2524f6-d551-4b6d-be5d-6aac50aa1a44 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.118258] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1142.118258] env[62558]: value = "task-1267401" [ 1142.118258] env[62558]: _type = "Task" [ 1142.118258] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.126378] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267401, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.225340] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.227692] env[62558]: DEBUG nova.compute.manager [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1142.228806] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e145ca4d-5ba5-4119-a604-eea0c24d42a3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.521048] env[62558]: INFO nova.compute.manager [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Took 0.54 seconds to detach 1 volumes for instance. [ 1142.523204] env[62558]: DEBUG nova.compute.manager [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Deleting volume: 393854fc-2de5-43ff-8f13-b76ec330d217 {{(pid=62558) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1142.628184] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267401, 'name': ReconfigVM_Task, 'duration_secs': 0.287712} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.628657] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfigured VM instance instance-0000006f to attach disk [datastore1] f6d46374-7a19-4079-8257-1ac69ac82b81/f6d46374-7a19-4079-8257-1ac69ac82b81.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1142.629378] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e98746cd-005e-460e-a845-e77c238a999c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.634968] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1142.634968] env[62558]: value = "task-1267403" [ 1142.634968] env[62558]: _type = "Task" [ 1142.634968] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.643286] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267403, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.060895] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.061325] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.061460] env[62558]: DEBUG nova.objects.instance [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'resources' on Instance uuid 5be1007a-6229-4d45-82d0-53d81a28204c {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.144342] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267403, 'name': Rename_Task, 'duration_secs': 0.130258} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.144604] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1143.144842] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-458e8642-881d-4a2f-b912-55e37eee019a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.151151] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1143.151151] env[62558]: value = "task-1267404" [ 1143.151151] env[62558]: _type = "Task" [ 1143.151151] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.158058] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.243896] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecf632f-f3f7-440b-a0a4-bcef37348b1d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.251186] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Doing hard reboot of VM {{(pid=62558) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1143.251436] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-fe5c1ae5-0d2c-4fb6-91ec-cf4a1a8baac5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.256762] env[62558]: DEBUG oslo_vmware.api [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1143.256762] env[62558]: value = "task-1267405" [ 1143.256762] env[62558]: _type = "Task" [ 1143.256762] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.263987] env[62558]: DEBUG oslo_vmware.api [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267405, 'name': ResetVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.637081] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69d47d1-d771-4ef7-8aaf-1c9bc2bce063 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.644553] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076bef95-86d4-40cc-91ad-f18fea62aa39 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.678283] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8423de86-25f4-4e65-a6e7-adf4daf431d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.685187] env[62558]: DEBUG oslo_vmware.api [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267404, 'name': PowerOnVM_Task, 'duration_secs': 0.411767} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.687094] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1143.687340] env[62558]: INFO nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Took 6.35 seconds to spawn the instance on the hypervisor. [ 1143.687532] env[62558]: DEBUG nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1143.688303] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff52b0d-9909-40de-aa31-718e8391cdd9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.691456] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771ca25e-3d7d-4333-9b2c-90e4faa628d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.704456] env[62558]: DEBUG nova.compute.provider_tree [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.766334] env[62558]: DEBUG oslo_vmware.api [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267405, 'name': ResetVM_Task, 'duration_secs': 0.093861} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.766614] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Did hard reboot of VM {{(pid=62558) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1143.766790] env[62558]: DEBUG nova.compute.manager [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1143.767584] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568a565e-cc51-4d37-b312-7170ea50b7dc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.211349] env[62558]: DEBUG nova.scheduler.client.report [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.221026] env[62558]: INFO nova.compute.manager [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Took 11.07 seconds to build instance. [ 1144.278413] env[62558]: DEBUG oslo_concurrency.lockutils [None req-81576b53-74c5-4cd1-82a0-06a45bc2b0ef tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.782s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.722547] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.725021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-de27a5fe-1267-4e50-b63b-1d123bb8b583 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.587s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.741646] env[62558]: INFO nova.scheduler.client.report [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocations for instance 5be1007a-6229-4d45-82d0-53d81a28204c [ 1145.165217] env[62558]: DEBUG nova.compute.manager [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Received event network-changed-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1145.165474] env[62558]: DEBUG nova.compute.manager [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Refreshing instance network info cache due to event network-changed-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1145.165706] env[62558]: DEBUG oslo_concurrency.lockutils [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] Acquiring lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.165865] env[62558]: DEBUG oslo_concurrency.lockutils [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] Acquired lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.166197] env[62558]: DEBUG nova.network.neutron [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Refreshing network info cache for port eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1145.249201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-635116b5-6e67-4506-8a87-e259f4b03c98 tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5be1007a-6229-4d45-82d0-53d81a28204c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.560s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.879995] env[62558]: DEBUG nova.network.neutron [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updated VIF entry in instance network info cache for port eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1145.880397] env[62558]: DEBUG nova.network.neutron [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating instance_info_cache with network_info: [{"id": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "address": "fa:16:3e:2f:d9:bf", "network": {"id": "89535486-8a79-4936-8104-dd4a9e174713", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1086750906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "46aa5789fccf4925aa49f9c4050c8463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359c2c31-99c4-41d7-a513-3bc4825897a0", "external-id": "nsx-vlan-transportzone-173", "segmentation_id": 173, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeaa9fa2a-20", "ovs_interfaceid": "eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.383821] env[62558]: DEBUG oslo_concurrency.lockutils [req-9dc9eee1-e85c-4226-a68d-d202db04e3fc req-d059c19c-187d-454e-a496-df3d5b818946 service nova] Releasing lock "refresh_cache-f6d46374-7a19-4079-8257-1ac69ac82b81" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.471856] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.472098] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.472320] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.472516] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.472734] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.474948] env[62558]: INFO nova.compute.manager [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Terminating instance [ 1146.476738] env[62558]: DEBUG nova.compute.manager [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1146.476936] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1146.477829] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a375e64e-2812-44b0-b209-7c2118e28cdf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.485883] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.486145] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3fba6b94-81b0-4101-b53b-4c1029a13879 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.493016] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1146.493016] env[62558]: value = "task-1267406" [ 1146.493016] env[62558]: _type = "Task" [ 1146.493016] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.501275] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.003133] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267406, 'name': PowerOffVM_Task, 'duration_secs': 0.44676} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.003423] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1147.003596] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1147.003844] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c54fce5-c066-446e-bd1e-8d80b0ee8ce7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.059060] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1147.059285] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1147.059470] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore2] 5082012b-e8ed-41d6-a036-b64cbb1e363c {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.059736] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a46fa05-1134-440d-a860-e720e1324e43 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.065652] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1147.065652] env[62558]: value = "task-1267408" [ 1147.065652] env[62558]: _type = "Task" [ 1147.065652] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.072799] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.576359] env[62558]: DEBUG oslo_vmware.api [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239738} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.576724] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.576780] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1147.576956] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1147.577150] env[62558]: INFO nova.compute.manager [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1147.577443] env[62558]: DEBUG oslo.service.loopingcall [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.577691] env[62558]: DEBUG nova.compute.manager [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1147.577790] env[62558]: DEBUG nova.network.neutron [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1147.819459] env[62558]: DEBUG nova.compute.manager [req-8f83897e-d1a3-4441-a4d5-b5ace7ceff62 req-ff5539c8-1bfc-4344-82dd-b635436c6040 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Received event network-vif-deleted-fbef2b6a-ad6d-410d-9495-f3c25b3d937a {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.819736] env[62558]: INFO nova.compute.manager [req-8f83897e-d1a3-4441-a4d5-b5ace7ceff62 req-ff5539c8-1bfc-4344-82dd-b635436c6040 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Neutron deleted interface fbef2b6a-ad6d-410d-9495-f3c25b3d937a; detaching it from the instance and deleting it from the info cache [ 1147.819889] env[62558]: DEBUG nova.network.neutron [req-8f83897e-d1a3-4441-a4d5-b5ace7ceff62 req-ff5539c8-1bfc-4344-82dd-b635436c6040 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.299735] env[62558]: DEBUG nova.network.neutron [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.322730] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-995c5eae-e04b-47c2-9de2-8df67d267e68 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.332556] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a274a245-0802-48c3-a191-68c697f8f5d3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.357731] env[62558]: DEBUG nova.compute.manager [req-8f83897e-d1a3-4441-a4d5-b5ace7ceff62 req-ff5539c8-1bfc-4344-82dd-b635436c6040 service nova] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Detach interface failed, port_id=fbef2b6a-ad6d-410d-9495-f3c25b3d937a, reason: Instance 5082012b-e8ed-41d6-a036-b64cbb1e363c could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1148.802391] env[62558]: INFO nova.compute.manager [-] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Took 1.22 seconds to deallocate network for instance. [ 1149.309121] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.309431] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.309676] env[62558]: DEBUG nova.objects.instance [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'resources' on Instance uuid 5082012b-e8ed-41d6-a036-b64cbb1e363c {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.878893] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a207f179-058c-4ab8-baee-f623782c648a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.886771] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bd90cd-18c3-4cbc-8755-0b260efc7e5d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.915304] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88dd28cd-1411-40b5-a7fc-aab5808f3bfb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.921938] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d699f2f7-0265-4e0f-92a1-77825e3f9844 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.935321] env[62558]: DEBUG nova.compute.provider_tree [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.439115] env[62558]: DEBUG nova.scheduler.client.report [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.944314] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.962927] env[62558]: INFO nova.scheduler.client.report [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocations for instance 5082012b-e8ed-41d6-a036-b64cbb1e363c [ 1151.472647] env[62558]: DEBUG oslo_concurrency.lockutils [None req-273e15c2-8a6e-45d8-9a02-c6b111d909bb tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "5082012b-e8ed-41d6-a036-b64cbb1e363c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.046410] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.046824] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.046960] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.047169] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.047443] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.049959] env[62558]: INFO nova.compute.manager [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Terminating instance [ 1155.051802] env[62558]: DEBUG nova.compute.manager [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1155.051982] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1155.052825] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053a63dd-9c48-4873-aabf-2dd966128c00 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.060547] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.060812] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8653e726-8c2c-4059-805b-ad2d3cfb4a42 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.066911] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1155.066911] env[62558]: value = "task-1267410" [ 1155.066911] env[62558]: _type = "Task" [ 1155.066911] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.074225] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.577080] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267410, 'name': PowerOffVM_Task, 'duration_secs': 0.178646} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.577388] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1155.577562] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1155.577811] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfb943cf-cfec-4659-acfd-88fd3b670c99 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.634077] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1155.634305] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1155.634482] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleting the datastore file [datastore2] 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1155.634742] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75d9d7bc-0afa-4427-ba46-4a8641adb6fc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.640953] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for the task: (returnval){ [ 1155.640953] env[62558]: value = "task-1267412" [ 1155.640953] env[62558]: _type = "Task" [ 1155.640953] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.648115] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267412, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.150362] env[62558]: DEBUG oslo_vmware.api [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Task: {'id': task-1267412, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126825} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.150761] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.150804] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1156.150973] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1156.151172] env[62558]: INFO nova.compute.manager [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1156.151417] env[62558]: DEBUG oslo.service.loopingcall [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.151615] env[62558]: DEBUG nova.compute.manager [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1156.151710] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1156.396279] env[62558]: DEBUG nova.compute.manager [req-53c7d58e-a7f3-4618-b5c3-377792588a5b req-021939d4-bfa6-479d-a716-aa4bfa1ea2f3 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Received event network-vif-deleted-6fcb24d4-f48e-4628-995d-10f53f01154b {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1156.396446] env[62558]: INFO nova.compute.manager [req-53c7d58e-a7f3-4618-b5c3-377792588a5b req-021939d4-bfa6-479d-a716-aa4bfa1ea2f3 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Neutron deleted interface 6fcb24d4-f48e-4628-995d-10f53f01154b; detaching it from the instance and deleting it from the info cache [ 1156.396628] env[62558]: DEBUG nova.network.neutron [req-53c7d58e-a7f3-4618-b5c3-377792588a5b req-021939d4-bfa6-479d-a716-aa4bfa1ea2f3 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.876261] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.899264] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5005d2c-1369-4031-bb55-5fd2d5fe1816 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.908999] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44930bd0-bcf9-430a-aab5-99688e60f5c4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.932024] env[62558]: DEBUG nova.compute.manager [req-53c7d58e-a7f3-4618-b5c3-377792588a5b req-021939d4-bfa6-479d-a716-aa4bfa1ea2f3 service nova] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Detach interface failed, port_id=6fcb24d4-f48e-4628-995d-10f53f01154b, reason: Instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1157.379558] env[62558]: INFO nova.compute.manager [-] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Took 1.23 seconds to deallocate network for instance. [ 1157.885861] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.886181] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.886415] env[62558]: DEBUG nova.objects.instance [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lazy-loading 'resources' on Instance uuid 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.440339] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046675cb-b942-4562-bcb6-5db72f85da35 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.447611] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315a199b-709a-4533-806a-c1818a319116 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.476604] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3070b4-fbec-4cfd-bcb6-9551b74b3a0b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.482953] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afe2738-a37d-46dd-82e5-da7a8efda45d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.496425] env[62558]: DEBUG nova.compute.provider_tree [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.999496] env[62558]: DEBUG nova.scheduler.client.report [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1159.504265] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.524190] env[62558]: INFO nova.scheduler.client.report [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Deleted allocations for instance 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b [ 1160.032317] env[62558]: DEBUG oslo_concurrency.lockutils [None req-96e9a015-5dca-4ede-b882-9a07e8207cdc tempest-ServerActionsTestOtherA-1771351709 tempest-ServerActionsTestOtherA-1771351709-project-member] Lock "4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.985s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.369893] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "7c3e82e2-1d08-4985-9b55-de845550c845" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.370149] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.872066] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1182.394814] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.395110] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.396628] env[62558]: INFO nova.compute.claims [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1183.448618] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b9fcbe-c0d5-469a-8781-3ae1c218e435 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.456213] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225eee05-aaf8-4219-b6b8-4d152e04390f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.134078] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb37013-3f1a-4e95-b09d-41d4b978cff3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.138585] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.138816] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.143152] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33416dd8-64a7-49bf-b0dd-4c6532a9202d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.156784] env[62558]: DEBUG nova.compute.provider_tree [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.630910] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.631299] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1184.643348] env[62558]: DEBUG nova.compute.utils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1184.659181] env[62558]: DEBUG nova.scheduler.client.report [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1185.146677] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.163187] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.768s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.163648] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1185.630718] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.630903] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1185.631052] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1185.667636] env[62558]: DEBUG nova.compute.utils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1185.668877] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1185.669080] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1185.716517] env[62558]: DEBUG nova.policy [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '985bef0c1e7e429e986753953b597926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '688bb982a71a4730aa1652ea837e1b73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1185.770338] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.770632] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.770891] env[62558]: INFO nova.compute.manager [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Attaching volume 183e03c5-d0b6-422d-942c-5bf03130f69c to /dev/sdb [ 1185.803053] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521a474b-a86a-42e8-b392-21d94bf92ba9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.810459] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1950ab3-5dce-43f8-9b1e-d891db007d33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.823264] env[62558]: DEBUG nova.virt.block_device [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating existing volume attachment record: fa4c0371-6e50-420a-bb30-1892af332ac7 {{(pid=62558) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1185.959086] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Successfully created port: f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1186.134659] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Skipping network cache update for instance because it is Building. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1186.171739] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1186.175602] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1186.175805] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.176070] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1186.176333] env[62558]: DEBUG nova.objects.instance [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lazy-loading 'info_cache' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.184728] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1187.211818] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1187.212093] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1187.212263] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.212448] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1187.212598] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.212749] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1187.212958] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1187.213429] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1187.213750] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1187.213935] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1187.214147] env[62558]: DEBUG nova.virt.hardware [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1187.215014] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67df7101-bbdb-4c90-971a-ef60fc929413 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.223226] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ce1a87-07ec-4b2e-a040-c282973e8ced {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.340692] env[62558]: DEBUG nova.compute.manager [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Received event network-vif-plugged-f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.340692] env[62558]: DEBUG oslo_concurrency.lockutils [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] Acquiring lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.340692] env[62558]: DEBUG oslo_concurrency.lockutils [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] Lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.340692] env[62558]: DEBUG oslo_concurrency.lockutils [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] Lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.340692] env[62558]: DEBUG nova.compute.manager [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] No waiting events found dispatching network-vif-plugged-f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1187.340692] env[62558]: WARNING nova.compute.manager [req-1deea7e2-1591-41ed-b8cf-efdb6adfa1ce req-6e16d97d-7cd5-43fb-ad34-6545db75b6cb service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Received unexpected event network-vif-plugged-f10324d1-f03e-4d7a-a290-5ff75658c2a1 for instance with vm_state building and task_state spawning. [ 1187.418990] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Successfully updated port: f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1187.905522] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.921033] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.921174] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.921306] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1188.408303] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.408580] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1188.408732] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.451767] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1188.565072] env[62558]: DEBUG nova.network.neutron [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Updating instance_info_cache with network_info: [{"id": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "address": "fa:16:3e:3a:01:fd", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf10324d1-f0", "ovs_interfaceid": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.630417] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.630609] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.067874] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.068296] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance network_info: |[{"id": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "address": "fa:16:3e:3a:01:fd", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf10324d1-f0", "ovs_interfaceid": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1189.068751] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:01:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f10324d1-f03e-4d7a-a290-5ff75658c2a1', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1189.076376] env[62558]: DEBUG oslo.service.loopingcall [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1189.076589] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1189.076809] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-772c60d0-e451-4ee9-9a2d-7a32f845b0ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.097083] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1189.097083] env[62558]: value = "task-1267415" [ 1189.097083] env[62558]: _type = "Task" [ 1189.097083] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.104096] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267415, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.365716] env[62558]: DEBUG nova.compute.manager [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Received event network-changed-f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.365970] env[62558]: DEBUG nova.compute.manager [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Refreshing instance network info cache due to event network-changed-f10324d1-f03e-4d7a-a290-5ff75658c2a1. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1189.366389] env[62558]: DEBUG oslo_concurrency.lockutils [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] Acquiring lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.366576] env[62558]: DEBUG oslo_concurrency.lockutils [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] Acquired lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.366758] env[62558]: DEBUG nova.network.neutron [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Refreshing network info cache for port f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1189.607191] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267415, 'name': CreateVM_Task, 'duration_secs': 0.273691} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.607534] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1189.608087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.608260] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.608597] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1189.608852] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a982168-ce98-4829-bfdb-d151234b5953 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.612874] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1189.612874] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526c487e-b1fc-a5fb-6e6d-7c98939e6c74" [ 1189.612874] env[62558]: _type = "Task" [ 1189.612874] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.620327] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526c487e-b1fc-a5fb-6e6d-7c98939e6c74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.041178] env[62558]: DEBUG nova.network.neutron [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Updated VIF entry in instance network info cache for port f10324d1-f03e-4d7a-a290-5ff75658c2a1. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1190.041636] env[62558]: DEBUG nova.network.neutron [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Updating instance_info_cache with network_info: [{"id": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "address": "fa:16:3e:3a:01:fd", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf10324d1-f0", "ovs_interfaceid": "f10324d1-f03e-4d7a-a290-5ff75658c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.127382] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]526c487e-b1fc-a5fb-6e6d-7c98939e6c74, 'name': SearchDatastore_Task, 'duration_secs': 0.008749} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.127899] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.128343] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1190.128726] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.129038] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.129366] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1190.129746] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aff8dc2e-9fbd-4ea4-a7d3-5555d656da57 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.138703] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1190.139020] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1190.140108] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2115647-488a-4a74-8cdb-7e55aba287b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.146618] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1190.146618] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fab046-a7d6-98a6-3918-172f66181ad8" [ 1190.146618] env[62558]: _type = "Task" [ 1190.146618] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.158281] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52fab046-a7d6-98a6-3918-172f66181ad8, 'name': SearchDatastore_Task, 'duration_secs': 0.007639} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.159382] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dab7b83-e0fd-41f9-83a2-32a8e6af14db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.165506] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1190.165506] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aef1b4-fa99-4d46-249e-dcb80b3bd573" [ 1190.165506] env[62558]: _type = "Task" [ 1190.165506] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.176130] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aef1b4-fa99-4d46-249e-dcb80b3bd573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.365965] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Volume attach. Driver type: vmdk {{(pid=62558) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1190.366290] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272628', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'name': 'volume-183e03c5-d0b6-422d-942c-5bf03130f69c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f6d46374-7a19-4079-8257-1ac69ac82b81', 'attached_at': '', 'detached_at': '', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'serial': '183e03c5-d0b6-422d-942c-5bf03130f69c'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1190.367159] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cd29a2-2886-4a68-a446-0621312c236a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.384152] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48d14e8-d1a7-4555-9ea3-b4446d32841f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.407494] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-183e03c5-d0b6-422d-942c-5bf03130f69c/volume-183e03c5-d0b6-422d-942c-5bf03130f69c.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1190.407712] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c5507ba-bddb-40d5-9409-b44d480645a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.424393] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1190.424393] env[62558]: value = "task-1267416" [ 1190.424393] env[62558]: _type = "Task" [ 1190.424393] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.431469] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.544371] env[62558]: DEBUG oslo_concurrency.lockutils [req-368e65c1-6a5e-4c9c-9b9b-3e4587859e16 req-dde794ac-0a30-4f4b-8ee0-cb201b5ee366 service nova] Releasing lock "refresh_cache-7c3e82e2-1d08-4985-9b55-de845550c845" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.630301] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.630677] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.630677] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.676379] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52aef1b4-fa99-4d46-249e-dcb80b3bd573, 'name': SearchDatastore_Task, 'duration_secs': 0.007922} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.676640] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.676941] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1190.677248] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-37679f5b-386d-41df-b029-1f9bebe08697 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.683676] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1190.683676] env[62558]: value = "task-1267417" [ 1190.683676] env[62558]: _type = "Task" [ 1190.683676] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.691781] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.934891] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267416, 'name': ReconfigVM_Task, 'duration_secs': 0.321751} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.935191] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-183e03c5-d0b6-422d-942c-5bf03130f69c/volume-183e03c5-d0b6-422d-942c-5bf03130f69c.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1190.940270] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c0c57d-e133-424e-a485-c25a9df6a337 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.958991] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1190.958991] env[62558]: value = "task-1267418" [ 1190.958991] env[62558]: _type = "Task" [ 1190.958991] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.967834] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267418, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.135070] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.135070] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.135070] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.135070] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1191.136031] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9dcf3f7-dfb0-4de5-ade7-0dc531f8c735 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.144194] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92da431-127b-43bf-9494-91cb7e6847e4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.157716] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3265855-459f-4cdc-af5b-92b1b33256bf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.164130] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d780786-8a24-4fc6-8295-f2f18e963b25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.193058] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180740MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1191.193197] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.193349] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.202380] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45957} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.202637] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1191.202851] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1191.203110] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-815f3f22-5dc4-4a1b-8c6e-d896ec51e0ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.208728] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1191.208728] env[62558]: value = "task-1267419" [ 1191.208728] env[62558]: _type = "Task" [ 1191.208728] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.215894] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.468613] env[62558]: DEBUG oslo_vmware.api [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267418, 'name': ReconfigVM_Task, 'duration_secs': 0.202865} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.468922] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272628', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'name': 'volume-183e03c5-d0b6-422d-942c-5bf03130f69c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f6d46374-7a19-4079-8257-1ac69ac82b81', 'attached_at': '', 'detached_at': '', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'serial': '183e03c5-d0b6-422d-942c-5bf03130f69c'} {{(pid=62558) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1191.719416] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058989} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.719799] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1191.720514] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c139f97-64b0-4d12-9b80-69faa973e7e3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.742638] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1191.742990] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e99e6fa-65eb-44a1-8397-fb4446bfe686 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.763857] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1191.763857] env[62558]: value = "task-1267420" [ 1191.763857] env[62558]: _type = "Task" [ 1191.763857] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.772410] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.222079] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 4e26d956-f0f4-4ab1-b23c-fd6859875929 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.222257] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance f6d46374-7a19-4079-8257-1ac69ac82b81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.222408] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 7c3e82e2-1d08-4985-9b55-de845550c845 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.222575] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1192.222714] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1192.273764] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.275471] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9d02f7-dbd9-438f-a132-6d3ab2fd5e00 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.282653] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0331407-47e7-49cf-ac84-3b52c0394499 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.313764] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b684bc-5810-4281-a793-e5913dae5b54 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.324175] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac83225-a2d5-4af4-940b-e5b4c02665d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.344080] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.509254] env[62558]: DEBUG nova.objects.instance [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid f6d46374-7a19-4079-8257-1ac69ac82b81 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.774984] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267420, 'name': ReconfigVM_Task, 'duration_secs': 0.5745} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.775361] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1192.775785] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-449c98c4-cfb7-4911-93ca-8dfa9e868c9c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.782963] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1192.782963] env[62558]: value = "task-1267421" [ 1192.782963] env[62558]: _type = "Task" [ 1192.782963] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.790481] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267421, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.847086] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.015733] env[62558]: DEBUG oslo_concurrency.lockutils [None req-3444c02c-4ed2-4cf5-859c-69edf686b4a2 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.245s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.216567] env[62558]: DEBUG oslo_concurrency.lockutils [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.216759] env[62558]: DEBUG oslo_concurrency.lockutils [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.292150] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267421, 'name': Rename_Task, 'duration_secs': 0.154067} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.292421] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1193.292654] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d0f95fa-439a-4391-b0dd-f6767c3557fb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.298340] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1193.298340] env[62558]: value = "task-1267422" [ 1193.298340] env[62558]: _type = "Task" [ 1193.298340] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.305269] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.352199] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1193.352400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.159s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.722145] env[62558]: INFO nova.compute.manager [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Detaching volume 183e03c5-d0b6-422d-942c-5bf03130f69c [ 1193.753924] env[62558]: INFO nova.virt.block_device [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Attempting to driver detach volume 183e03c5-d0b6-422d-942c-5bf03130f69c from mountpoint /dev/sdb [ 1193.754182] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Volume detach. Driver type: vmdk {{(pid=62558) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1193.754375] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272628', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'name': 'volume-183e03c5-d0b6-422d-942c-5bf03130f69c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f6d46374-7a19-4079-8257-1ac69ac82b81', 'attached_at': '', 'detached_at': '', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'serial': '183e03c5-d0b6-422d-942c-5bf03130f69c'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1193.755238] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913d8188-913b-4f8e-bd57-a08c82dbe8c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.776511] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3021a3b2-4f97-4e95-a734-487519d894e7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.783116] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f378b54c-ccf2-4349-995e-d21e21e6f55f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.805078] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4268e83b-f788-485e-8b69-85750dd1c87d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.811410] env[62558]: DEBUG oslo_vmware.api [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267422, 'name': PowerOnVM_Task, 'duration_secs': 0.43079} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.821150] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1193.821337] env[62558]: INFO nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1193.821530] env[62558]: DEBUG nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1193.821817] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] The volume has not been displaced from its original location: [datastore2] volume-183e03c5-d0b6-422d-942c-5bf03130f69c/volume-183e03c5-d0b6-422d-942c-5bf03130f69c.vmdk. No consolidation needed. {{(pid=62558) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1193.826846] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1193.827549] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79bc0dd-37b6-4844-808b-78abb4dda1c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.829744] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e951352c-771c-42af-98a0-9251f9309a33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.848570] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1193.848570] env[62558]: value = "task-1267423" [ 1193.848570] env[62558]: _type = "Task" [ 1193.848570] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.856444] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267423, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.361617] env[62558]: INFO nova.compute.manager [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Took 11.98 seconds to build instance. [ 1194.365676] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267423, 'name': ReconfigVM_Task, 'duration_secs': 0.213296} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.365926] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1194.370864] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8599007-cc85-4e57-9337-cb254b68f8e1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.386343] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1194.386343] env[62558]: value = "task-1267424" [ 1194.386343] env[62558]: _type = "Task" [ 1194.386343] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.394064] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267424, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.863906] env[62558]: DEBUG oslo_concurrency.lockutils [None req-4cd569ee-bb1c-4a69-8d20-8a16415b07e5 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.494s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1194.896232] env[62558]: DEBUG oslo_vmware.api [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267424, 'name': ReconfigVM_Task, 'duration_secs': 0.131197} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.896519] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-272628', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'name': 'volume-183e03c5-d0b6-422d-942c-5bf03130f69c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f6d46374-7a19-4079-8257-1ac69ac82b81', 'attached_at': '', 'detached_at': '', 'volume_id': '183e03c5-d0b6-422d-942c-5bf03130f69c', 'serial': '183e03c5-d0b6-422d-942c-5bf03130f69c'} {{(pid=62558) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1195.435436] env[62558]: DEBUG nova.objects.instance [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'flavor' on Instance uuid f6d46374-7a19-4079-8257-1ac69ac82b81 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.790547] env[62558]: INFO nova.compute.manager [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Rebuilding instance [ 1195.829042] env[62558]: DEBUG nova.compute.manager [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1195.829905] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d0db74-dc95-44e2-a9c9-6582b4fdc1a2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.340984] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.341443] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce39891a-78c0-4ed6-90a4-6d1f0b7956cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.348400] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.348708] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1196.348708] env[62558]: value = "task-1267425" [ 1196.348708] env[62558]: _type = "Task" [ 1196.348708] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.356617] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.442978] env[62558]: DEBUG oslo_concurrency.lockutils [None req-658c9fc1-6936-4c12-9d68-13b914883ba9 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.226s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.858071] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267425, 'name': PowerOffVM_Task, 'duration_secs': 0.204106} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.858294] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.858605] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1196.859425] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f725aa32-a1e3-4fd3-8835-bf506d6d91b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.865862] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.866173] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79306198-7577-40d5-9b4f-163f36edab85 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.920432] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.920663] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.920841] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleting the datastore file [datastore2] 7c3e82e2-1d08-4985-9b55-de845550c845 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.921131] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffb8896a-570f-493e-9b2d-1d58c179fb7f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.926875] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1196.926875] env[62558]: value = "task-1267427" [ 1196.926875] env[62558]: _type = "Task" [ 1196.926875] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.934225] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267427, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.436253] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267427, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144541} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.437367] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1197.437367] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1197.437367] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1197.468846] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.469100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.469364] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.469579] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.469756] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.471913] env[62558]: INFO nova.compute.manager [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Terminating instance [ 1197.473611] env[62558]: DEBUG nova.compute.manager [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1197.473799] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1197.474793] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c6b86b-bdc8-4d11-be39-a3fe3dbbf3f5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.481669] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1197.481893] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c422df12-e955-4dda-a8f3-4f2b3ca8761c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.488072] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1197.488072] env[62558]: value = "task-1267428" [ 1197.488072] env[62558]: _type = "Task" [ 1197.488072] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.495557] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267428, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.998153] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267428, 'name': PowerOffVM_Task, 'duration_secs': 0.172539} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.998473] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1197.998672] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1197.998927] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b52f1558-af36-4218-82f3-0be6890a4cbb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.058893] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1198.059122] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1198.059325] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleting the datastore file [datastore1] f6d46374-7a19-4079-8257-1ac69ac82b81 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.059589] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6eddf5dc-490c-4501-8adb-7e9c1340dd33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.065684] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for the task: (returnval){ [ 1198.065684] env[62558]: value = "task-1267430" [ 1198.065684] env[62558]: _type = "Task" [ 1198.065684] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.072951] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.472862] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1198.473199] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1198.473372] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1198.473565] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1198.473718] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1198.473870] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1198.474094] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1198.474262] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1198.474439] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1198.474602] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1198.474774] env[62558]: DEBUG nova.virt.hardware [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1198.475655] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dfa990-3750-4b48-9f74-d6162c0a5faf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.482657] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dc8cbc-eca9-486c-9596-ec40287aa337 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.495311] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:01:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f10324d1-f03e-4d7a-a290-5ff75658c2a1', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1198.502421] env[62558]: DEBUG oslo.service.loopingcall [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.502648] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1198.502847] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-def15ef7-0098-4643-b3c9-4c1f8ee875ad {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.522322] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1198.522322] env[62558]: value = "task-1267431" [ 1198.522322] env[62558]: _type = "Task" [ 1198.522322] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.529082] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267431, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.575380] env[62558]: DEBUG oslo_vmware.api [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Task: {'id': task-1267430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116656} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.575671] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.575907] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1198.576146] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1198.576358] env[62558]: INFO nova.compute.manager [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1198.576609] env[62558]: DEBUG oslo.service.loopingcall [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.576825] env[62558]: DEBUG nova.compute.manager [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.576940] env[62558]: DEBUG nova.network.neutron [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1199.029970] env[62558]: DEBUG nova.compute.manager [req-bdef7ac6-b354-4bd6-83a8-ed01c4562cfb req-3ff75569-fb80-46b6-90e7-c4be767f76c5 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Received event network-vif-deleted-eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.030362] env[62558]: INFO nova.compute.manager [req-bdef7ac6-b354-4bd6-83a8-ed01c4562cfb req-3ff75569-fb80-46b6-90e7-c4be767f76c5 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Neutron deleted interface eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c; detaching it from the instance and deleting it from the info cache [ 1199.030553] env[62558]: DEBUG nova.network.neutron [req-bdef7ac6-b354-4bd6-83a8-ed01c4562cfb req-3ff75569-fb80-46b6-90e7-c4be767f76c5 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.037710] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267431, 'name': CreateVM_Task, 'duration_secs': 0.296619} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.037962] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1199.038569] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.038732] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.039344] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1199.039701] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f35c54c4-bb4e-4428-896e-211ac6ea775e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.046792] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1199.046792] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1a328-d6c0-f74e-790c-ceab9bfa2a38" [ 1199.046792] env[62558]: _type = "Task" [ 1199.046792] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.055806] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1a328-d6c0-f74e-790c-ceab9bfa2a38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.506451] env[62558]: DEBUG nova.network.neutron [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.533073] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adc27ee1-4df7-46e8-bb29-7db351d3827a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.542368] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06592e97-984c-42ea-a78d-251a93f5b339 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.560767] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52a1a328-d6c0-f74e-790c-ceab9bfa2a38, 'name': SearchDatastore_Task, 'duration_secs': 0.009195} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.561087] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.561317] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.561553] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.561704] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.561882] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.562139] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c88a45b-299d-45da-9a4b-39e8b875be6e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.570987] env[62558]: DEBUG nova.compute.manager [req-bdef7ac6-b354-4bd6-83a8-ed01c4562cfb req-3ff75569-fb80-46b6-90e7-c4be767f76c5 service nova] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Detach interface failed, port_id=eaa9fa2a-2056-4a74-a3ec-6ee0f122c61c, reason: Instance f6d46374-7a19-4079-8257-1ac69ac82b81 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1199.573649] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.573649] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.573649] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de62515e-42f9-4b41-8720-8173d86e7b79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.577603] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1199.577603] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb266f-cc2b-f978-0e96-b6bc67f8c6a9" [ 1199.577603] env[62558]: _type = "Task" [ 1199.577603] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.585072] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb266f-cc2b-f978-0e96-b6bc67f8c6a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.010352] env[62558]: INFO nova.compute.manager [-] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Took 1.43 seconds to deallocate network for instance. [ 1200.087788] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52bb266f-cc2b-f978-0e96-b6bc67f8c6a9, 'name': SearchDatastore_Task, 'duration_secs': 0.00873} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.088508] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20c447f3-940e-4833-8388-f8730a9b8c6b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.093707] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1200.093707] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527342ce-356d-e95b-3d70-40e9a68c5db3" [ 1200.093707] env[62558]: _type = "Task" [ 1200.093707] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.101225] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527342ce-356d-e95b-3d70-40e9a68c5db3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.517080] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.517392] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.517618] env[62558]: DEBUG nova.objects.instance [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lazy-loading 'resources' on Instance uuid f6d46374-7a19-4079-8257-1ac69ac82b81 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1200.604694] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]527342ce-356d-e95b-3d70-40e9a68c5db3, 'name': SearchDatastore_Task, 'duration_secs': 0.009155} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.604934] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.605209] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1200.605458] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57eaeb7d-1f48-43c5-ac2b-76ee89f66d21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.611248] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1200.611248] env[62558]: value = "task-1267432" [ 1200.611248] env[62558]: _type = "Task" [ 1200.611248] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.618117] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.076168] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b74de43-0951-4f14-80e7-2b3542e1dcb2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.083335] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578628cc-55e6-44b6-ae8a-4b01618fd0d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.112301] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67589453-070b-4df0-ad78-c7d2247ee3b6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.124456] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cecda6-75bc-435a-af56-b4470700a490 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.128056] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42936} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.128339] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore1] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1201.128611] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1201.129128] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-643ba30a-5aaa-4b7a-9865-a03530e18643 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.138228] env[62558]: DEBUG nova.compute.provider_tree [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.143338] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1201.143338] env[62558]: value = "task-1267433" [ 1201.143338] env[62558]: _type = "Task" [ 1201.143338] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.150875] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267433, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.626394] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.642011] env[62558]: DEBUG nova.scheduler.client.report [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.655777] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267433, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05841} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.656069] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1201.656810] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562b1879-6a3a-4614-b91f-122db8e93fe3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.679229] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.679685] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-895674ab-8700-4fa2-a607-7b83df1018f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.698035] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1201.698035] env[62558]: value = "task-1267434" [ 1201.698035] env[62558]: _type = "Task" [ 1201.698035] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.705236] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267434, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.151617] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.173264] env[62558]: INFO nova.scheduler.client.report [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Deleted allocations for instance f6d46374-7a19-4079-8257-1ac69ac82b81 [ 1202.207513] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267434, 'name': ReconfigVM_Task, 'duration_secs': 0.272079} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.207840] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 7c3e82e2-1d08-4985-9b55-de845550c845/7c3e82e2-1d08-4985-9b55-de845550c845.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.208512] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf671847-10a3-4c99-b1de-494d3f7d346d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.215179] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1202.215179] env[62558]: value = "task-1267435" [ 1202.215179] env[62558]: _type = "Task" [ 1202.215179] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.223850] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267435, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.682102] env[62558]: DEBUG oslo_concurrency.lockutils [None req-98021aa0-352c-43a2-a3cc-a63f74a1e9c3 tempest-AttachVolumeNegativeTest-1398038761 tempest-AttachVolumeNegativeTest-1398038761-project-member] Lock "f6d46374-7a19-4079-8257-1ac69ac82b81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.213s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.724597] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267435, 'name': Rename_Task, 'duration_secs': 0.144292} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.724870] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1202.725141] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-274e65dd-a6ce-4673-bcc7-0451c6122f40 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.730441] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1202.730441] env[62558]: value = "task-1267436" [ 1202.730441] env[62558]: _type = "Task" [ 1202.730441] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.737438] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.239725] env[62558]: DEBUG oslo_vmware.api [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267436, 'name': PowerOnVM_Task, 'duration_secs': 0.422445} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.241050] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1203.241050] env[62558]: DEBUG nova.compute.manager [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1203.241200] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c038879-2894-4351-8e89-76d6dcb394c5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.756410] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.756766] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.757097] env[62558]: DEBUG nova.objects.instance [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62558) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1204.765164] env[62558]: DEBUG oslo_concurrency.lockutils [None req-abecbc90-a04f-4bd5-bb81-25c0fe073b21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.384355] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "7c3e82e2-1d08-4985-9b55-de845550c845" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.384528] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.384747] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.384930] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.385133] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.387363] env[62558]: INFO nova.compute.manager [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Terminating instance [ 1205.389177] env[62558]: DEBUG nova.compute.manager [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1205.389382] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1205.390218] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f520dbd9-6a9a-46f0-9dd5-31f3bb54bf4a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.397961] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.398209] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-273de774-52fe-4fd6-bd11-a31875553983 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.404247] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1205.404247] env[62558]: value = "task-1267438" [ 1205.404247] env[62558]: _type = "Task" [ 1205.404247] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.411673] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.914638] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267438, 'name': PowerOffVM_Task, 'duration_secs': 0.172976} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.915042] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1205.915105] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.915337] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccbdb441-332e-4bb5-b262-9aa7b3fc4c03 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.976867] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1205.977058] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Deleting contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1205.977263] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleting the datastore file [datastore1] 7c3e82e2-1d08-4985-9b55-de845550c845 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.977532] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5dfdfcf7-1166-4ab7-87b4-82e17cfb58cb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.985516] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1205.985516] env[62558]: value = "task-1267440" [ 1205.985516] env[62558]: _type = "Task" [ 1205.985516] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.993044] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267440, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.494803] env[62558]: DEBUG oslo_vmware.api [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267440, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176866} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.495765] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.495765] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Deleted contents of the VM from datastore datastore1 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.495765] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1206.495931] env[62558]: INFO nova.compute.manager [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1206.496125] env[62558]: DEBUG oslo.service.loopingcall [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1206.496335] env[62558]: DEBUG nova.compute.manager [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1206.496420] env[62558]: DEBUG nova.network.neutron [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1206.741276] env[62558]: DEBUG nova.compute.manager [req-5d396f58-fa2c-4e13-aa15-290016739339 req-f60063c0-2212-4cef-a031-bdff01956ac4 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Received event network-vif-deleted-f10324d1-f03e-4d7a-a290-5ff75658c2a1 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1206.741594] env[62558]: INFO nova.compute.manager [req-5d396f58-fa2c-4e13-aa15-290016739339 req-f60063c0-2212-4cef-a031-bdff01956ac4 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Neutron deleted interface f10324d1-f03e-4d7a-a290-5ff75658c2a1; detaching it from the instance and deleting it from the info cache [ 1206.741845] env[62558]: DEBUG nova.network.neutron [req-5d396f58-fa2c-4e13-aa15-290016739339 req-f60063c0-2212-4cef-a031-bdff01956ac4 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.221236] env[62558]: DEBUG nova.network.neutron [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.244603] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4eaefc24-430f-403c-a7f6-e64d259eb2a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.253555] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74f28e1-45f1-4a3b-85a5-820c5ab70593 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.275315] env[62558]: DEBUG nova.compute.manager [req-5d396f58-fa2c-4e13-aa15-290016739339 req-f60063c0-2212-4cef-a031-bdff01956ac4 service nova] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Detach interface failed, port_id=f10324d1-f03e-4d7a-a290-5ff75658c2a1, reason: Instance 7c3e82e2-1d08-4985-9b55-de845550c845 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1207.722876] env[62558]: INFO nova.compute.manager [-] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Took 1.23 seconds to deallocate network for instance. [ 1208.229877] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.230194] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.230435] env[62558]: DEBUG nova.objects.instance [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'resources' on Instance uuid 7c3e82e2-1d08-4985-9b55-de845550c845 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.776249] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcff5e30-884c-4e92-ad14-cc1b945ac54e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.783265] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be837fb-13ed-4149-95d0-c8a965f2cf76 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.812444] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b7b947-e955-488d-abe2-bbe092f4dd36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.818915] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27293e48-572b-423f-9db2-06fa92286712 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.831245] env[62558]: DEBUG nova.compute.provider_tree [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.333827] env[62558]: DEBUG nova.scheduler.client.report [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1209.839804] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.609s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.858553] env[62558]: INFO nova.scheduler.client.report [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted allocations for instance 7c3e82e2-1d08-4985-9b55-de845550c845 [ 1210.367244] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e4eeee6b-98d1-4e54-b959-b20f136ce711 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "7c3e82e2-1d08-4985-9b55-de845550c845" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.983s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.163308] env[62558]: DEBUG nova.compute.manager [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1212.680984] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.681292] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.185727] env[62558]: INFO nova.compute.claims [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1213.692494] env[62558]: INFO nova.compute.resource_tracker [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating resource usage from migration 0114fb0c-03ac-4b27-a22e-de68806bb7f9 [ 1213.738559] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919f86a1-977b-482c-9458-bf202b257db8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.746283] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6657a758-b747-406e-b9fc-3a20667ba374 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.776218] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b861a1-2f4b-483d-b281-31c08036270a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.783189] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5da92d-b491-4e81-80f1-1b89e574a90a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.795908] env[62558]: DEBUG nova.compute.provider_tree [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.298954] env[62558]: DEBUG nova.scheduler.client.report [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1214.804290] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.123s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.804515] env[62558]: INFO nova.compute.manager [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Migrating [ 1215.319802] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.320218] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.320307] env[62558]: DEBUG nova.network.neutron [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1216.013607] env[62558]: DEBUG nova.network.neutron [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.516286] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.031618] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bf2227-e525-4cd4-81ca-2f69a02d40c1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.050851] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1218.556920] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1218.557269] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d2ca9cf-ada2-4912-8dd4-41dfa7f202e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.564053] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1218.564053] env[62558]: value = "task-1267443" [ 1218.564053] env[62558]: _type = "Task" [ 1218.564053] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.571301] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.074184] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267443, 'name': PowerOffVM_Task, 'duration_secs': 0.178377} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.074600] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1219.074646] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1219.581139] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1219.581407] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1219.581569] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1219.581754] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1219.581906] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1219.582066] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1219.582279] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1219.582442] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1219.582613] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1219.582780] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1219.582954] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1219.587886] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fe9a3f6-2588-4370-9bb7-c0db3d91a656 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.603722] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1219.603722] env[62558]: value = "task-1267444" [ 1219.603722] env[62558]: _type = "Task" [ 1219.603722] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.611502] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267444, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.113580] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267444, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.614388] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267444, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.115047] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267444, 'name': ReconfigVM_Task, 'duration_secs': 1.107502} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.115429] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1221.621794] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1221.622034] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1221.622198] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1221.622355] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1221.622508] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1221.622661] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1221.622972] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1221.623166] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1221.623345] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1221.623508] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1221.623685] env[62558]: DEBUG nova.virt.hardware [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1221.628936] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1221.629231] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-247cc02d-e662-42e7-a880-c6bf85caa713 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.647409] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1221.647409] env[62558]: value = "task-1267445" [ 1221.647409] env[62558]: _type = "Task" [ 1221.647409] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.654555] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267445, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.157220] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267445, 'name': ReconfigVM_Task, 'duration_secs': 0.154493} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.157592] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1222.158377] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c179c6be-73b5-4819-8e03-22a357d3a1e9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.180219] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1222.180752] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c11fa073-fc77-4d0b-9e70-4f79977328e5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.197494] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1222.197494] env[62558]: value = "task-1267446" [ 1222.197494] env[62558]: _type = "Task" [ 1222.197494] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.204382] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267446, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.707836] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267446, 'name': ReconfigVM_Task, 'duration_secs': 0.241192} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.708226] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929/4e26d956-f0f4-4ab1-b23c-fd6859875929.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1222.708530] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1223.215499] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e851c9e6-1b4b-4edd-b8fe-ce2d61b1e67a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.234988] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cc6e93-ccf2-47d9-8fe9-fb7d30d7ed4b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.252951] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1223.791822] env[62558]: DEBUG nova.network.neutron [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Port 44bbeb25-dd9d-4cc1-914a-a3edad861b31 binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1224.813928] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.814343] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.814343] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.847442] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.847707] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.847812] env[62558]: DEBUG nova.network.neutron [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1226.549366] env[62558]: DEBUG nova.network.neutron [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.052054] env[62558]: DEBUG oslo_concurrency.lockutils [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.577230] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f15e8a-5348-4e02-baae-e383813b6c0c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.595410] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fad258-5c73-4890-ba0d-6cdf05276d75 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.601884] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1228.108534] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1228.108946] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-749bc903-0348-43a3-adcc-fba49bdc6c05 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.116011] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1228.116011] env[62558]: value = "task-1267447" [ 1228.116011] env[62558]: _type = "Task" [ 1228.116011] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.123459] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.626105] env[62558]: DEBUG oslo_vmware.api [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267447, 'name': PowerOnVM_Task, 'duration_secs': 0.359316} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.626458] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1228.626565] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-9874eb65-561b-40d1-bfe4-134af39fda73 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance '4e26d956-f0f4-4ab1-b23c-fd6859875929' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1230.456777] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.457120] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.457297] env[62558]: DEBUG nova.compute.manager [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Going to confirm migration 5 {{(pid=62558) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1231.026151] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1231.026360] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.026580] env[62558]: DEBUG nova.network.neutron [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1231.026780] env[62558]: DEBUG nova.objects.instance [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'info_cache' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.224832] env[62558]: DEBUG nova.network.neutron [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [{"id": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "address": "fa:16:3e:e1:0e:3f", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44bbeb25-dd", "ovs_interfaceid": "44bbeb25-dd9d-4cc1-914a-a3edad861b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.727657] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-4e26d956-f0f4-4ab1-b23c-fd6859875929" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.727951] env[62558]: DEBUG nova.objects.instance [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'migration_context' on Instance uuid 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.231715] env[62558]: DEBUG nova.objects.base [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Object Instance<4e26d956-f0f4-4ab1-b23c-fd6859875929> lazy-loaded attributes: info_cache,migration_context {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1233.232715] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4546de01-7502-403a-a705-9e9d599eb3ca {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.251562] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c62e91fc-f506-495b-a97a-8ae3ae45ee1c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.256503] env[62558]: DEBUG oslo_vmware.api [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1233.256503] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ca6a49-550c-fa6e-28ce-8bc1245012ac" [ 1233.256503] env[62558]: _type = "Task" [ 1233.256503] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.263607] env[62558]: DEBUG oslo_vmware.api [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ca6a49-550c-fa6e-28ce-8bc1245012ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.766705] env[62558]: DEBUG oslo_vmware.api [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ca6a49-550c-fa6e-28ce-8bc1245012ac, 'name': SearchDatastore_Task, 'duration_secs': 0.006726} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.767077] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.767278] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.310409] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75806d93-df79-4fc0-a3ce-b5978b243a9f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.317649] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959d85f4-45d4-41ea-a580-24cc62eee55b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.345700] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c790e0d-0a51-47cc-ac8d-b31d62f9d3ae {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.352114] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24383e3a-488d-4cee-b0a1-1ca943edbdbf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.364473] env[62558]: DEBUG nova.compute.provider_tree [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1234.867429] env[62558]: DEBUG nova.scheduler.client.report [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1235.878017] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.110s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.433680] env[62558]: INFO nova.scheduler.client.report [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted allocation for migration 0114fb0c-03ac-4b27-a22e-de68806bb7f9 [ 1236.940069] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e6a48d34-aa83-491a-b2eb-af6144aba4a6 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.482s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.460021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.460369] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.460586] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.460793] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.460968] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.463266] env[62558]: INFO nova.compute.manager [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Terminating instance [ 1237.465178] env[62558]: DEBUG nova.compute.manager [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1237.465372] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1237.466211] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f35d256-f42f-4127-915b-3db9d703ca48 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.475167] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1237.475258] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1bafcb5-b750-4d46-b0b7-54842125c6a8 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.481063] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1237.481063] env[62558]: value = "task-1267448" [ 1237.481063] env[62558]: _type = "Task" [ 1237.481063] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.488640] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267448, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.991480] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267448, 'name': PowerOffVM_Task, 'duration_secs': 0.172155} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.991829] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1237.991885] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1237.992121] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-575291f5-fecf-46ec-b109-40aeb3f6fe79 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.048359] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1238.048593] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1238.048756] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleting the datastore file [datastore2] 4e26d956-f0f4-4ab1-b23c-fd6859875929 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1238.049060] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47ef111f-debf-4c6f-a3ed-0579084d7be5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.055284] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1238.055284] env[62558]: value = "task-1267450" [ 1238.055284] env[62558]: _type = "Task" [ 1238.055284] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.063119] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.565725] env[62558]: DEBUG oslo_vmware.api [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130368} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.565996] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1238.566189] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1238.566376] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1238.566553] env[62558]: INFO nova.compute.manager [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1238.566800] env[62558]: DEBUG oslo.service.loopingcall [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1238.566999] env[62558]: DEBUG nova.compute.manager [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1238.567111] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1238.989706] env[62558]: DEBUG nova.compute.manager [req-6cb1f5c3-bc45-432d-b166-a1f97b37c12f req-180c0d09-e0ce-4df5-b4b7-9499feabd9ca service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Received event network-vif-deleted-44bbeb25-dd9d-4cc1-914a-a3edad861b31 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1238.989925] env[62558]: INFO nova.compute.manager [req-6cb1f5c3-bc45-432d-b166-a1f97b37c12f req-180c0d09-e0ce-4df5-b4b7-9499feabd9ca service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Neutron deleted interface 44bbeb25-dd9d-4cc1-914a-a3edad861b31; detaching it from the instance and deleting it from the info cache [ 1238.990142] env[62558]: DEBUG nova.network.neutron [req-6cb1f5c3-bc45-432d-b166-a1f97b37c12f req-180c0d09-e0ce-4df5-b4b7-9499feabd9ca service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.464559] env[62558]: DEBUG nova.network.neutron [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.492464] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a7c05a0-c258-4934-b932-26305cbd629d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.503932] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedc25d0-4b03-467e-a0d5-7b6359c8ac2b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.527347] env[62558]: DEBUG nova.compute.manager [req-6cb1f5c3-bc45-432d-b166-a1f97b37c12f req-180c0d09-e0ce-4df5-b4b7-9499feabd9ca service nova] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Detach interface failed, port_id=44bbeb25-dd9d-4cc1-914a-a3edad861b31, reason: Instance 4e26d956-f0f4-4ab1-b23c-fd6859875929 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1239.968108] env[62558]: INFO nova.compute.manager [-] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Took 1.40 seconds to deallocate network for instance. [ 1240.473838] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.474217] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.474336] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.496506] env[62558]: INFO nova.scheduler.client.report [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted allocations for instance 4e26d956-f0f4-4ab1-b23c-fd6859875929 [ 1241.004267] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e8c21c62-2638-4f7d-89f8-8427b6ccc47c tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "4e26d956-f0f4-4ab1-b23c-fd6859875929" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.544s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.411041] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "d11bf400-f80e-4815-9641-6af6bd1b4460" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.411322] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.913846] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1243.435253] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.435546] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.437045] env[62558]: INFO nova.compute.claims [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1244.473678] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c65ad0-0309-4635-88d7-b4b5da4b3161 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.480890] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ad9002-4475-4a23-8ba1-2df6c5b1f160 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.515684] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7628de03-ff24-4fbd-b8c5-a1265fd43ef6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.522937] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9becd2-b578-4dfc-bca6-194a89886925 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.535508] env[62558]: DEBUG nova.compute.provider_tree [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1245.038394] env[62558]: DEBUG nova.scheduler.client.report [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1245.543621] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.544176] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1245.630619] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1245.630832] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1246.049018] env[62558]: DEBUG nova.compute.utils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1246.050519] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1246.050693] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1246.100026] env[62558]: DEBUG nova.policy [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '985bef0c1e7e429e986753953b597926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '688bb982a71a4730aa1652ea837e1b73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1246.369447] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Successfully created port: 67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1246.553972] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1246.634667] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.564723] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1247.590024] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1247.590328] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1247.590507] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1247.590696] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1247.590846] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1247.591369] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1247.591369] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1247.591369] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1247.591542] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1247.591693] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1247.591868] env[62558]: DEBUG nova.virt.hardware [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1247.592734] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9634f0f-a7b9-4c50-bba9-a9781c7e11fa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.600389] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7011d433-f888-4eed-9491-d4b62551ba95 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.630573] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1247.630731] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1247.767520] env[62558]: DEBUG nova.compute.manager [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Received event network-vif-plugged-67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1247.767756] env[62558]: DEBUG oslo_concurrency.lockutils [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] Acquiring lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.767972] env[62558]: DEBUG oslo_concurrency.lockutils [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.768166] env[62558]: DEBUG oslo_concurrency.lockutils [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.768360] env[62558]: DEBUG nova.compute.manager [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] No waiting events found dispatching network-vif-plugged-67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1247.768505] env[62558]: WARNING nova.compute.manager [req-15f1e47a-05b6-4a78-8bc5-cc62f231b790 req-c4e13d2b-2dc9-48ba-aa05-3d850f0f834d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Received unexpected event network-vif-plugged-67a02dd8-d432-4ccf-9bb9-0120932ab67d for instance with vm_state building and task_state spawning. [ 1248.133775] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Didn't find any instances for network info cache update. {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1248.294221] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Successfully updated port: 67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1248.316336] env[62558]: DEBUG nova.compute.manager [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Received event network-changed-67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1248.316558] env[62558]: DEBUG nova.compute.manager [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Refreshing instance network info cache due to event network-changed-67a02dd8-d432-4ccf-9bb9-0120932ab67d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1248.316774] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.316947] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.317090] env[62558]: DEBUG nova.network.neutron [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Refreshing network info cache for port 67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1248.630494] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.797697] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.847324] env[62558]: DEBUG nova.network.neutron [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1248.920926] env[62558]: DEBUG nova.network.neutron [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.423572] env[62558]: DEBUG oslo_concurrency.lockutils [req-fd8f1955-afb9-464a-a35d-e4dec66c5b03 req-5a038fc2-9b8f-42f7-a3d5-2b484b5ab56d service nova] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.424184] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.424443] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1249.631543] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.954182] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1250.072556] env[62558]: DEBUG nova.network.neutron [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.575112] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.575434] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance network_info: |[{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1250.575988] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:0f:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '67a02dd8-d432-4ccf-9bb9-0120932ab67d', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1250.583481] env[62558]: DEBUG oslo.service.loopingcall [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1250.583685] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1250.583909] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-104fab61-9034-4b83-9d19-5739189f17d4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.603515] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1250.603515] env[62558]: value = "task-1267451" [ 1250.603515] env[62558]: _type = "Task" [ 1250.603515] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.610557] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267451, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.630069] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.113257] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267451, 'name': CreateVM_Task, 'duration_secs': 0.27826} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.113665] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1251.114100] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.114281] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.114607] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1251.114862] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db1d9b6c-f3c0-46c4-aac9-021411966481 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.119258] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1251.119258] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f2884-77af-8709-eded-df20d9847b88" [ 1251.119258] env[62558]: _type = "Task" [ 1251.119258] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.126293] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f2884-77af-8709-eded-df20d9847b88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.132201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.132400] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.132564] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1251.132712] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1251.133429] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e690a83c-a5e4-42ad-aad0-2529f7c1fc7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.139787] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63bb397-cadf-449d-9bc3-77c49814cd32 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.153517] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de2dc6e-e2b0-4bce-86b8-801df1fcde48 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.159419] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7969e2-1212-4e11-b733-e981716521b9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.187016] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181140MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1251.187150] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.187356] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1251.630271] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528f2884-77af-8709-eded-df20d9847b88, 'name': SearchDatastore_Task, 'duration_secs': 0.010096} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.630592] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.630830] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1251.631085] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.631241] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.631430] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1251.631689] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2da251a2-227f-433a-804d-9fb3809bae50 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.639584] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1251.639763] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1251.640494] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee2d4449-3562-413b-a370-9e458ef1038b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.645493] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1251.645493] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528a18f7-cc30-1211-29a9-d27c9dbdb585" [ 1251.645493] env[62558]: _type = "Task" [ 1251.645493] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.652814] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528a18f7-cc30-1211-29a9-d27c9dbdb585, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.155996] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528a18f7-cc30-1211-29a9-d27c9dbdb585, 'name': SearchDatastore_Task, 'duration_secs': 0.008128} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.156750] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba1470ef-8640-4eed-b2ee-d13749694b58 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.162055] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1252.162055] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52abf821-09ba-c120-5add-da733eb6a222" [ 1252.162055] env[62558]: _type = "Task" [ 1252.162055] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.170096] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52abf821-09ba-c120-5add-da733eb6a222, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.212941] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d11bf400-f80e-4815-9641-6af6bd1b4460 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1252.213193] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1252.213346] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1252.240780] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd883c51-31f0-40bc-8168-296ef0a34411 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.249653] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6b15cf-1d51-48d3-89cf-f1b62e280677 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.279252] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cec92a9-20ac-4e40-ab3e-534fdb9d0fbc {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.287227] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8b9a92-536c-4ce9-a3f3-cd0aaa9ba9b0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.300878] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.673237] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52abf821-09ba-c120-5add-da733eb6a222, 'name': SearchDatastore_Task, 'duration_secs': 0.009438} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.673638] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.673786] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1252.674084] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-885d605e-061b-4cc0-aa31-1b6816079419 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.682771] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1252.682771] env[62558]: value = "task-1267452" [ 1252.682771] env[62558]: _type = "Task" [ 1252.682771] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.690515] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.805076] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1253.192763] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444514} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.193165] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1253.193299] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1253.193506] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68a3f3a7-bdee-470b-af9b-98cfd5cd41b4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.199142] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1253.199142] env[62558]: value = "task-1267453" [ 1253.199142] env[62558]: _type = "Task" [ 1253.199142] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.207444] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.308922] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1253.309155] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.122s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.709841] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057721} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.710115] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1253.710885] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41714ea-acf6-422b-b154-1de67553d0d0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.732149] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1253.732345] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63db01a7-39bf-4b82-8430-a911b61dd71f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.751031] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1253.751031] env[62558]: value = "task-1267454" [ 1253.751031] env[62558]: _type = "Task" [ 1253.751031] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.758591] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.260895] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267454, 'name': ReconfigVM_Task, 'duration_secs': 0.265933} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.261372] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfigured VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1254.261809] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f28b08df-6c10-415a-b724-89fa82dbdc91 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.268895] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1254.268895] env[62558]: value = "task-1267455" [ 1254.268895] env[62558]: _type = "Task" [ 1254.268895] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.276113] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267455, 'name': Rename_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.309730] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1254.309924] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1254.778458] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267455, 'name': Rename_Task, 'duration_secs': 0.124051} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.778734] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1254.778979] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-febe80b7-dfcf-419f-a223-b1f5d0a69fb1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.784489] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1254.784489] env[62558]: value = "task-1267456" [ 1254.784489] env[62558]: _type = "Task" [ 1254.784489] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.791497] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267456, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.294525] env[62558]: DEBUG oslo_vmware.api [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267456, 'name': PowerOnVM_Task, 'duration_secs': 0.404792} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.294892] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1255.294999] env[62558]: INFO nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1255.295177] env[62558]: DEBUG nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1255.295967] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e8202b-001f-4242-8154-eb693f414b32 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.626579] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.812255] env[62558]: INFO nova.compute.manager [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Took 12.39 seconds to build instance. [ 1256.314201] env[62558]: DEBUG oslo_concurrency.lockutils [None req-090afffe-11c6-4a1a-9f56-b72a96e2c570 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.903s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.690574] env[62558]: DEBUG nova.compute.manager [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Received event network-changed-67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1256.690788] env[62558]: DEBUG nova.compute.manager [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Refreshing instance network info cache due to event network-changed-67a02dd8-d432-4ccf-9bb9-0120932ab67d. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1256.690988] env[62558]: DEBUG oslo_concurrency.lockutils [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.691164] env[62558]: DEBUG oslo_concurrency.lockutils [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.691333] env[62558]: DEBUG nova.network.neutron [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Refreshing network info cache for port 67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1257.401196] env[62558]: DEBUG nova.network.neutron [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updated VIF entry in instance network info cache for port 67a02dd8-d432-4ccf-9bb9-0120932ab67d. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1257.401586] env[62558]: DEBUG nova.network.neutron [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.904953] env[62558]: DEBUG oslo_concurrency.lockutils [req-d91c70c5-7cfe-4bd6-a1c1-4672083b6ba5 req-a2844237-c407-43b6-8efa-ecf12c5de7d6 service nova] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1294.144028] env[62558]: DEBUG nova.compute.manager [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Stashing vm_state: active {{(pid=62558) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1294.660538] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.660815] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.166565] env[62558]: INFO nova.compute.claims [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1295.673337] env[62558]: INFO nova.compute.resource_tracker [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating resource usage from migration 1a580cfe-bb2b-4644-9d99-4f3aff061364 [ 1295.708373] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c032c4-e773-452e-88d0-76e497c2607a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.716030] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210f85bb-7113-4ce9-a752-543480b40f58 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.746056] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d09b9f-df19-45f7-8906-17ea39e95cc0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.753070] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f901bd-c1fc-4d7d-918a-6ee4e8e86a7b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.765777] env[62558]: DEBUG nova.compute.provider_tree [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1296.268561] env[62558]: DEBUG nova.scheduler.client.report [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1296.773352] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.112s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.773687] env[62558]: INFO nova.compute.manager [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Migrating [ 1297.288432] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1297.288834] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.288918] env[62558]: DEBUG nova.network.neutron [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1297.995765] env[62558]: DEBUG nova.network.neutron [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.498701] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.014136] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d493ba7-c293-4a2a-b4ea-44a0fcd5e516 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.032407] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 0 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1300.538579] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.538973] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-838e61d0-2fe9-4e69-a1a4-afafc88dccf5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.547384] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1300.547384] env[62558]: value = "task-1267457" [ 1300.547384] env[62558]: _type = "Task" [ 1300.547384] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.555078] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.057707] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267457, 'name': PowerOffVM_Task, 'duration_secs': 0.171953} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.058089] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1301.058144] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 17 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1301.564840] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1301.565120] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1301.565289] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1301.565478] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1301.565630] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1301.565782] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1301.566026] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1301.566201] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1301.566400] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1301.566582] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1301.566765] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1301.571776] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5422b8e-77b3-4fe7-8871-81adeae4cac9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.587500] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1301.587500] env[62558]: value = "task-1267458" [ 1301.587500] env[62558]: _type = "Task" [ 1301.587500] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.596613] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267458, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.097021] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267458, 'name': ReconfigVM_Task, 'duration_secs': 0.159344} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.097398] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 33 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1302.603850] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1302.604137] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1302.604309] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1302.604505] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1302.604657] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1302.604812] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1302.605035] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1302.605209] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1302.605381] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1302.605551] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1302.605725] env[62558]: DEBUG nova.virt.hardware [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1302.611060] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1302.611363] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6f5a6d1-25c3-4aac-b558-daa845f903d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.629297] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1302.629297] env[62558]: value = "task-1267459" [ 1302.629297] env[62558]: _type = "Task" [ 1302.629297] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.636725] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267459, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.139348] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267459, 'name': ReconfigVM_Task, 'duration_secs': 0.14153} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.139709] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1303.140398] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c90e7f-f3b7-4b40-b5b7-d27d43c82bf6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.163711] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1303.164355] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60f94948-9fa4-43f0-ad9b-f3591cc9943a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.181599] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1303.181599] env[62558]: value = "task-1267460" [ 1303.181599] env[62558]: _type = "Task" [ 1303.181599] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.190424] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.691671] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267460, 'name': ReconfigVM_Task, 'duration_secs': 0.227926} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.691936] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfigured VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1303.692222] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 50 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1304.199167] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e94eca-0e26-47a3-98d2-a2b55f7ed4ce {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.217446] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de77b69-d7f8-4891-9df5-36b87da34e96 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.234203] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 67 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1304.772660] env[62558]: DEBUG nova.network.neutron [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Port 67a02dd8-d432-4ccf-9bb9-0120932ab67d binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1305.794092] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1305.794376] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1305.794500] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.828325] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1306.828556] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.828696] env[62558]: DEBUG nova.network.neutron [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1307.519850] env[62558]: DEBUG nova.network.neutron [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.630064] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1307.630257] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1308.022925] env[62558]: DEBUG oslo_concurrency.lockutils [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1308.545444] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6851477d-47e3-4f09-9be4-45cb81d9ec23 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.563780] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d973326-8850-40da-9b1d-a69efef7ceed {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.570238] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 83 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1308.630632] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1308.630783] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1308.630895] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1309.077359] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1309.077751] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0bb87c37-2ad9-4550-8b0c-de98b1006d35 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.085415] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1309.085415] env[62558]: value = "task-1267461" [ 1309.085415] env[62558]: _type = "Task" [ 1309.085415] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.093202] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.134315] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1309.134480] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.134601] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1309.134755] env[62558]: DEBUG nova.objects.instance [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lazy-loading 'info_cache' on Instance uuid d11bf400-f80e-4815-9641-6af6bd1b4460 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1309.595857] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267461, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.095846] env[62558]: DEBUG oslo_vmware.api [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267461, 'name': PowerOnVM_Task, 'duration_secs': 0.551567} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.096222] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1310.096365] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-74992435-78dc-4f38-814d-4d8795b9f127 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance 'd11bf400-f80e-4815-9641-6af6bd1b4460' progress to 100 {{(pid=62558) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1310.846673] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1311.349356] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1311.349768] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1311.349768] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.349915] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.350078] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.630452] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.630702] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.133723] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.133982] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.134089] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.134240] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1312.135178] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e876c4-8765-41d5-a957-6a0f3abb7b3f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.143629] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc16f4af-1a54-4a26-add3-9fea05c1d9b5 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.157881] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597c3e99-f293-4266-8228-2a0b911453cf {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.164690] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b361b47-4cc4-4b2f-93e3-51bed59f7014 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.194868] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181505MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1312.195096] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.195233] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.639887] env[62558]: DEBUG nova.network.neutron [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Port 67a02dd8-d432-4ccf-9bb9-0120932ab67d binding to destination host cpu-1 is already ACTIVE {{(pid=62558) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1312.640203] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1312.640361] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1312.640574] env[62558]: DEBUG nova.network.neutron [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1313.201208] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Applying migration context for instance d11bf400-f80e-4815-9641-6af6bd1b4460 as it has an incoming, in-progress migration 1a580cfe-bb2b-4644-9d99-4f3aff061364. Migration status is reverting {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1313.201820] env[62558]: INFO nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating resource usage from migration 1a580cfe-bb2b-4644-9d99-4f3aff061364 [ 1313.220868] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Migration 1a580cfe-bb2b-4644-9d99-4f3aff061364 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1313.221029] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance d11bf400-f80e-4815-9641-6af6bd1b4460 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1313.221217] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1313.221362] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=960MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1313.259545] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900865f6-aa4b-4320-8dc8-17e5297c2166 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.269234] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88ec04a-ebc7-4a6b-9c72-47a2677e75ef {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.300141] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1420a429-262b-479d-adeb-4e2402f5b04f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.307138] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a0c2f3-c8a5-4908-ba5f-db29ab024a2d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.320392] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1313.364318] env[62558]: DEBUG nova.network.neutron [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1313.823203] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1313.866286] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1314.327668] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1314.327889] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.133s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.369224] env[62558]: DEBUG nova.compute.manager [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62558) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1314.369463] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.369699] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.873133] env[62558]: DEBUG nova.objects.instance [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'migration_context' on Instance uuid d11bf400-f80e-4815-9641-6af6bd1b4460 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1315.328090] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.412608] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f0eae0-9643-45aa-aadf-996ebb9d0b36 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.419870] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157023ee-77e7-4509-acf0-7aa707bd2d25 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.448277] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfa04b7-6aab-42d6-9715-08717b204c52 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.454832] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e210e0a-9495-42f1-b445-98f93e3ccc6d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.468506] env[62558]: DEBUG nova.compute.provider_tree [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1315.625853] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.974310] env[62558]: DEBUG nova.scheduler.client.report [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1316.986746] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.617s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.521205] env[62558]: INFO nova.compute.manager [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Swapping old allocation on dict_keys(['dc830c09-1c36-446a-8af3-d3826bec8b3b']) held by migration 1a580cfe-bb2b-4644-9d99-4f3aff061364 for instance [ 1318.541290] env[62558]: DEBUG nova.scheduler.client.report [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Overwriting current allocation {'allocations': {'dc830c09-1c36-446a-8af3-d3826bec8b3b': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 149}}, 'project_id': '688bb982a71a4730aa1652ea837e1b73', 'user_id': '985bef0c1e7e429e986753953b597926', 'consumer_generation': 1} on consumer d11bf400-f80e-4815-9641-6af6bd1b4460 {{(pid=62558) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1318.621103] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1318.621312] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.621497] env[62558]: DEBUG nova.network.neutron [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1319.338580] env[62558]: DEBUG nova.network.neutron [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [{"id": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "address": "fa:16:3e:0d:0f:2e", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap67a02dd8-d4", "ovs_interfaceid": "67a02dd8-d432-4ccf-9bb9-0120932ab67d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1319.841229] env[62558]: DEBUG oslo_concurrency.lockutils [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-d11bf400-f80e-4815-9641-6af6bd1b4460" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1319.841718] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1319.842035] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d2be00e-02f3-4cb5-bef3-cbfb1f79259b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.849391] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1319.849391] env[62558]: value = "task-1267462" [ 1319.849391] env[62558]: _type = "Task" [ 1319.849391] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1319.857189] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.358989] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267462, 'name': PowerOffVM_Task, 'duration_secs': 0.222974} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.359286] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1320.359948] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1320.360185] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1320.360347] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1320.360533] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1320.360708] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1320.360870] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1320.361089] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1320.361259] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1320.361429] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1320.361605] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1320.361778] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1320.366554] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-914df176-04fe-4543-b101-0426b96766db {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.383172] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1320.383172] env[62558]: value = "task-1267463" [ 1320.383172] env[62558]: _type = "Task" [ 1320.383172] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.390244] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.893355] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267463, 'name': ReconfigVM_Task, 'duration_secs': 0.115854} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.894151] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c1161f-849b-4f28-8293-ce7c6be31687 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.911279] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1320.911524] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1320.911685] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1320.911869] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1320.912035] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1320.912191] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1320.912412] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1320.912575] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1320.912795] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1320.912907] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1320.913191] env[62558]: DEBUG nova.virt.hardware [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1320.913941] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa49f9e0-2137-42b6-844b-01d0ea372440 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.918982] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1320.918982] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5236bbdb-c3da-961a-9e97-3ab9009d74a1" [ 1320.918982] env[62558]: _type = "Task" [ 1320.918982] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.926571] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5236bbdb-c3da-961a-9e97-3ab9009d74a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.428704] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5236bbdb-c3da-961a-9e97-3ab9009d74a1, 'name': SearchDatastore_Task, 'duration_secs': 0.007057} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.433931] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1321.434225] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cb5d0ac-c9d8-4478-916c-3aab786d6bd2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.451682] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1321.451682] env[62558]: value = "task-1267464" [ 1321.451682] env[62558]: _type = "Task" [ 1321.451682] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.458769] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.961076] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267464, 'name': ReconfigVM_Task, 'duration_secs': 0.173934} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.961489] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=62558) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1321.962148] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26cafe7-a1e3-4728-9714-3367232d1354 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.982796] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1321.983039] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f5c2eb3-ccc6-45ad-a898-2468b98f5ef6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.000285] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1322.000285] env[62558]: value = "task-1267465" [ 1322.000285] env[62558]: _type = "Task" [ 1322.000285] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.007352] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267465, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.509835] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267465, 'name': ReconfigVM_Task, 'duration_secs': 0.258812} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.510131] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Reconfigured VM instance instance-00000071 to attach disk [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460/d11bf400-f80e-4815-9641-6af6bd1b4460.vmdk or device None with type thin {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1322.510969] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119ddcda-17b2-4e3f-ab68-0657729045f3 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.529239] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2711da9-7e21-40b7-ad59-c1863ff2e52a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.545899] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e615102-fbf9-44bf-9ca5-189b699c75ff {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.562333] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32633eed-c7a6-40de-b4ce-85a0f6906d4f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.568570] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1322.568795] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4fa3fc5-45ae-4b1b-86c7-e8376e2b1d9a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.574601] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1322.574601] env[62558]: value = "task-1267466" [ 1322.574601] env[62558]: _type = "Task" [ 1322.574601] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.581415] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.085512] env[62558]: DEBUG oslo_vmware.api [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267466, 'name': PowerOnVM_Task, 'duration_secs': 0.333926} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.085961] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1324.096767] env[62558]: INFO nova.compute.manager [None req-5840fa67-df92-47c5-a40a-cce311c71043 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance to original state: 'active' [ 1325.098250] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "d11bf400-f80e-4815-9641-6af6bd1b4460" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.098665] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.098804] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.099021] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.099206] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1325.101485] env[62558]: INFO nova.compute.manager [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Terminating instance [ 1325.103280] env[62558]: DEBUG nova.compute.manager [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1325.103479] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1325.104313] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16367995-4bcd-4ef9-a790-dbca37c23e4c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.112240] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1325.112744] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad8f2c4b-9d81-440f-9041-4e929492af26 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.118892] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1325.118892] env[62558]: value = "task-1267467" [ 1325.118892] env[62558]: _type = "Task" [ 1325.118892] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.129081] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.625436] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.629495] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267467, 'name': PowerOffVM_Task, 'duration_secs': 0.177024} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.629926] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1325.630119] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1325.630353] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-909e6a8b-bb7a-45b9-a0c9-73973ac9d3d2 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.689994] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1325.690234] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1325.690418] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleting the datastore file [datastore2] d11bf400-f80e-4815-9641-6af6bd1b4460 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1325.690686] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3789c355-0f16-4181-affe-779934765f21 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.696989] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1325.696989] env[62558]: value = "task-1267469" [ 1325.696989] env[62558]: _type = "Task" [ 1325.696989] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.704318] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.206916] env[62558]: DEBUG oslo_vmware.api [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141965} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.207263] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1326.207392] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1326.207569] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1326.207743] env[62558]: INFO nova.compute.manager [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1326.207989] env[62558]: DEBUG oslo.service.loopingcall [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1326.208201] env[62558]: DEBUG nova.compute.manager [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1326.208292] env[62558]: DEBUG nova.network.neutron [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1326.621151] env[62558]: DEBUG nova.compute.manager [req-3d7419d2-773f-4718-89b1-d3abe946056f req-bb28f7bd-aa13-4ef4-a88f-f72edd2fd788 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Received event network-vif-deleted-67a02dd8-d432-4ccf-9bb9-0120932ab67d {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1326.621228] env[62558]: INFO nova.compute.manager [req-3d7419d2-773f-4718-89b1-d3abe946056f req-bb28f7bd-aa13-4ef4-a88f-f72edd2fd788 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Neutron deleted interface 67a02dd8-d432-4ccf-9bb9-0120932ab67d; detaching it from the instance and deleting it from the info cache [ 1326.621349] env[62558]: DEBUG nova.network.neutron [req-3d7419d2-773f-4718-89b1-d3abe946056f req-bb28f7bd-aa13-4ef4-a88f-f72edd2fd788 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.101461] env[62558]: DEBUG nova.network.neutron [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.124839] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1b4a99e-3f4e-4a9b-8170-0beda6e35d0f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.134256] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a21f700-6275-4673-a6af-86720ac60ffa {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.155745] env[62558]: DEBUG nova.compute.manager [req-3d7419d2-773f-4718-89b1-d3abe946056f req-bb28f7bd-aa13-4ef4-a88f-f72edd2fd788 service nova] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Detach interface failed, port_id=67a02dd8-d432-4ccf-9bb9-0120932ab67d, reason: Instance d11bf400-f80e-4815-9641-6af6bd1b4460 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1327.604202] env[62558]: INFO nova.compute.manager [-] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Took 1.40 seconds to deallocate network for instance. [ 1328.112072] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1328.112519] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1328.112829] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1328.133794] env[62558]: INFO nova.scheduler.client.report [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted allocations for instance d11bf400-f80e-4815-9641-6af6bd1b4460 [ 1328.641342] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e512dce2-488c-4269-9760-4309610ee4dd tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "d11bf400-f80e-4815-9641-6af6bd1b4460" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.543s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1329.943456] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "9c56d217-086c-4869-baf9-0ec771294c32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.943814] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.446608] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Starting instance... {{(pid=62558) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1330.965868] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.966153] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.967667] env[62558]: INFO nova.compute.claims [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1332.001530] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec76f98-4a21-4f15-a20e-bd024b10d6d6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.008772] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab14330-b13b-4a59-b893-feaa7b489d38 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.037330] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09e1df1-da20-4637-a71d-cbd281f22938 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.043770] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f00acc-2105-4bf0-a973-9cf686e588a9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.057093] env[62558]: DEBUG nova.compute.provider_tree [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1332.559712] env[62558]: DEBUG nova.scheduler.client.report [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1333.064532] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.098s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1333.065102] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Start building networks asynchronously for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1333.570080] env[62558]: DEBUG nova.compute.utils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Using /dev/sd instead of None {{(pid=62558) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1333.571586] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Allocating IP information in the background. {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1333.572014] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] allocate_for_instance() {{(pid=62558) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1333.620358] env[62558]: DEBUG nova.policy [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '985bef0c1e7e429e986753953b597926', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '688bb982a71a4730aa1652ea837e1b73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62558) authorize /opt/stack/nova/nova/policy.py:201}} [ 1333.907325] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Successfully created port: e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1334.076020] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Start building block device mappings for instance. {{(pid=62558) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1335.086771] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Start spawning the instance on the hypervisor. {{(pid=62558) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1335.112542] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-11T04:58:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-11T04:58:05Z,direct_url=,disk_format='vmdk',id=bd09177c-5fd7-495c-893b-5ec219b6b3de,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b74813cb82fc4367b249c8af6705c455',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-11T04:58:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1335.112787] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1335.112951] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image limits 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1335.113152] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Flavor pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1335.113305] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Image pref 0:0:0 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1335.113457] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62558) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1335.113665] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1335.113828] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1335.114014] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Got 1 possible topologies {{(pid=62558) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1335.114184] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1335.114363] env[62558]: DEBUG nova.virt.hardware [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62558) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1335.115263] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c11f8c1-52fa-4f94-b5fb-4c1a2c29759b {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.122887] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d5cab2-a134-4a62-8778-e43d7a2ceb0e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.272930] env[62558]: DEBUG nova.compute.manager [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Received event network-vif-plugged-e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1335.273141] env[62558]: DEBUG oslo_concurrency.lockutils [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] Acquiring lock "9c56d217-086c-4869-baf9-0ec771294c32-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1335.273367] env[62558]: DEBUG oslo_concurrency.lockutils [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] Lock "9c56d217-086c-4869-baf9-0ec771294c32-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1335.273541] env[62558]: DEBUG oslo_concurrency.lockutils [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] Lock "9c56d217-086c-4869-baf9-0ec771294c32-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.273715] env[62558]: DEBUG nova.compute.manager [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] No waiting events found dispatching network-vif-plugged-e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1335.274055] env[62558]: WARNING nova.compute.manager [req-97ceca16-a7df-4ae3-adc8-81e05da39972 req-a13fd67b-d023-46d0-b475-d1f9c9e8ee8a service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Received unexpected event network-vif-plugged-e858943c-a264-4787-b618-3e9d78a3b6b5 for instance with vm_state building and task_state spawning. [ 1335.354058] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Successfully updated port: e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1335.857057] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1335.857228] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1335.857312] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1336.388655] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Instance cache missing network info. {{(pid=62558) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1336.514237] env[62558]: DEBUG nova.network.neutron [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.017335] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1337.017660] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Instance network_info: |[{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62558) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1337.018126] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:76:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e858943c-a264-4787-b618-3e9d78a3b6b5', 'vif_model': 'vmxnet3'}] {{(pid=62558) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1337.025563] env[62558]: DEBUG oslo.service.loopingcall [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1337.025776] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Creating VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1337.026011] env[62558]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bf68068-f93e-40e7-88a8-ff8a60b19554 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.046518] env[62558]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1337.046518] env[62558]: value = "task-1267470" [ 1337.046518] env[62558]: _type = "Task" [ 1337.046518] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.053808] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267470, 'name': CreateVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.299213] env[62558]: DEBUG nova.compute.manager [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Received event network-changed-e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1337.299213] env[62558]: DEBUG nova.compute.manager [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Refreshing instance network info cache due to event network-changed-e858943c-a264-4787-b618-3e9d78a3b6b5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1337.299213] env[62558]: DEBUG oslo_concurrency.lockutils [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.299347] env[62558]: DEBUG oslo_concurrency.lockutils [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.299425] env[62558]: DEBUG nova.network.neutron [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Refreshing network info cache for port e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1337.556566] env[62558]: DEBUG oslo_vmware.api [-] Task: {'id': task-1267470, 'name': CreateVM_Task, 'duration_secs': 0.284867} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.556888] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Created VM on the ESX host {{(pid=62558) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1337.557385] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.557562] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.557877] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1337.558140] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b1d6653-e12b-42cb-8118-022bdb8092ba {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.562538] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1337.562538] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5230b3d5-24ae-6166-4bba-90edd9e3df69" [ 1337.562538] env[62558]: _type = "Task" [ 1337.562538] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.569547] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5230b3d5-24ae-6166-4bba-90edd9e3df69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.973067] env[62558]: DEBUG nova.network.neutron [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updated VIF entry in instance network info cache for port e858943c-a264-4787-b618-3e9d78a3b6b5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1337.973899] env[62558]: DEBUG nova.network.neutron [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.072610] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]5230b3d5-24ae-6166-4bba-90edd9e3df69, 'name': SearchDatastore_Task, 'duration_secs': 0.009327} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.072903] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.073155] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Processing image bd09177c-5fd7-495c-893b-5ec219b6b3de {{(pid=62558) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1338.073420] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1338.073575] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.073757] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1338.074022] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b15831f-e12c-46e2-adc2-927061a61040 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.081490] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62558) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1338.081630] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62558) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1338.082322] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6d6bf47-c454-4438-af4d-15e0ceddef37 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.087338] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1338.087338] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528097cf-321e-d2dd-c05f-d2f433568e18" [ 1338.087338] env[62558]: _type = "Task" [ 1338.087338] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.094244] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528097cf-321e-d2dd-c05f-d2f433568e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.476805] env[62558]: DEBUG oslo_concurrency.lockutils [req-75cd37cf-8482-4d01-972a-06c704b28cd9 req-201bbb39-189f-4200-9c67-abf6222cdc0e service nova] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.597346] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]528097cf-321e-d2dd-c05f-d2f433568e18, 'name': SearchDatastore_Task, 'duration_secs': 0.007749} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.598112] env[62558]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5ace912-f2a9-4c4d-b7a1-dffb05589fa0 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.602723] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1338.602723] env[62558]: value = "session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff431d-b7ef-af48-a74b-dd9b6feb5f9d" [ 1338.602723] env[62558]: _type = "Task" [ 1338.602723] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.609550] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff431d-b7ef-af48-a74b-dd9b6feb5f9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.113311] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': session[523ad9a9-f1a6-0bc7-a77d-d3bc2d8e1350]52ff431d-b7ef-af48-a74b-dd9b6feb5f9d, 'name': SearchDatastore_Task, 'duration_secs': 0.00854} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.113608] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1339.113870] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9c56d217-086c-4869-baf9-0ec771294c32/9c56d217-086c-4869-baf9-0ec771294c32.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1339.114145] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7d23f48-9cc2-4908-8c51-da08825088a6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.120337] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1339.120337] env[62558]: value = "task-1267471" [ 1339.120337] env[62558]: _type = "Task" [ 1339.120337] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.127313] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.630689] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267471, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.370884} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.631107] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bd09177c-5fd7-495c-893b-5ec219b6b3de/bd09177c-5fd7-495c-893b-5ec219b6b3de.vmdk to [datastore2] 9c56d217-086c-4869-baf9-0ec771294c32/9c56d217-086c-4869-baf9-0ec771294c32.vmdk {{(pid=62558) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1339.631207] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Extending root virtual disk to 1048576 {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1339.631447] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6965d58b-f370-4bbf-850c-6a33b869402a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.636892] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1339.636892] env[62558]: value = "task-1267472" [ 1339.636892] env[62558]: _type = "Task" [ 1339.636892] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.643681] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267472, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.146223] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267472, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061236} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.146509] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Extended root virtual disk {{(pid=62558) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1340.147328] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8f407e-0ca1-4f5a-b17e-7d34900476d9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.169442] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 9c56d217-086c-4869-baf9-0ec771294c32/9c56d217-086c-4869-baf9-0ec771294c32.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1340.169442] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd2e4caa-4634-4fd2-9d02-2d3949cc58f7 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.188068] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1340.188068] env[62558]: value = "task-1267473" [ 1340.188068] env[62558]: _type = "Task" [ 1340.188068] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.195448] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267473, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.697865] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267473, 'name': ReconfigVM_Task, 'duration_secs': 0.259018} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.698241] env[62558]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 9c56d217-086c-4869-baf9-0ec771294c32/9c56d217-086c-4869-baf9-0ec771294c32.vmdk or device None with type sparse {{(pid=62558) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1340.698806] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b95fa1bb-50f7-422d-ac31-b93db3aac131 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.705246] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1340.705246] env[62558]: value = "task-1267474" [ 1340.705246] env[62558]: _type = "Task" [ 1340.705246] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.712757] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267474, 'name': Rename_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.215344] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267474, 'name': Rename_Task, 'duration_secs': 0.129089} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.215623] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1341.215869] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ca88410-7f78-41e2-adf4-1dfd9b78952c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.221760] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1341.221760] env[62558]: value = "task-1267475" [ 1341.221760] env[62558]: _type = "Task" [ 1341.221760] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.228679] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.730986] env[62558]: DEBUG oslo_vmware.api [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267475, 'name': PowerOnVM_Task, 'duration_secs': 0.464489} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.734091] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1341.734091] env[62558]: INFO nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1341.734091] env[62558]: DEBUG nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1341.734091] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910b5ebe-901f-4a1b-b95c-c284c1dd6b1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.249321] env[62558]: INFO nova.compute.manager [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Took 11.30 seconds to build instance. [ 1342.751876] env[62558]: DEBUG oslo_concurrency.lockutils [None req-b6902684-dd00-4596-aabe-06c45e5fbac8 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.808s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.947751] env[62558]: DEBUG nova.compute.manager [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Received event network-changed-e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1342.947915] env[62558]: DEBUG nova.compute.manager [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Refreshing instance network info cache due to event network-changed-e858943c-a264-4787-b618-3e9d78a3b6b5. {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1342.948152] env[62558]: DEBUG oslo_concurrency.lockutils [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1342.948303] env[62558]: DEBUG oslo_concurrency.lockutils [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.948469] env[62558]: DEBUG nova.network.neutron [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Refreshing network info cache for port e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1343.650849] env[62558]: DEBUG nova.network.neutron [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updated VIF entry in instance network info cache for port e858943c-a264-4787-b618-3e9d78a3b6b5. {{(pid=62558) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1343.651228] env[62558]: DEBUG nova.network.neutron [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.154370] env[62558]: DEBUG oslo_concurrency.lockutils [req-92397136-607e-4bca-9585-bad4b9bde147 req-341f90f0-cbaf-41ca-8e15-de6b31a29477 service nova] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1368.630460] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.630871] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62558) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1369.630626] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1370.630231] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1370.630405] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Starting heal instance info cache {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1370.630495] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Rebuilding the list of instances to heal {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1371.160824] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1371.161052] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1371.161135] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Forcefully refreshing network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1371.161291] env[62558]: DEBUG nova.objects.instance [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lazy-loading 'info_cache' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1372.903891] env[62558]: DEBUG nova.network.neutron [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1373.407014] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1373.407265] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updated the network info_cache for instance {{(pid=62558) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1373.407464] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.407644] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.407828] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.407974] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.408138] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.911881] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.912169] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1373.912361] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.912549] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62558) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1373.913478] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6382d07b-f954-44c6-82f0-d3cb268fb501 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.921808] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5dfa35-a3da-4b54-b7dc-13becb7b23a4 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.935426] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd9be04-bdb3-497a-9b8e-377f0b0e6e46 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.941314] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837b5d62-706d-4236-8119-8522be8d7df9 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.969483] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181498MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62558) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1373.969540] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1373.969719] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1374.993279] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Instance 9c56d217-086c-4869-baf9-0ec771294c32 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62558) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1374.993594] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1374.993654] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62558) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1375.009185] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing inventories for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1375.023480] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating ProviderTree inventory for provider dc830c09-1c36-446a-8af3-d3826bec8b3b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1375.023637] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Updating inventory in ProviderTree for provider dc830c09-1c36-446a-8af3-d3826bec8b3b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1375.033654] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing aggregate associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, aggregates: None {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1375.050465] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Refreshing trait associations for resource provider dc830c09-1c36-446a-8af3-d3826bec8b3b, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62558) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1375.074967] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3427dbf6-dcae-4f2d-be0e-e9d753129528 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.082340] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fef002b-1219-448a-b2e5-56acebc540bb {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.111660] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c290c9-36ce-494a-b1d8-fa3d5502cb63 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.118162] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ac718d-ba68-4f36-a5a7-c823cf579b33 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.131618] env[62558]: DEBUG nova.compute.provider_tree [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1375.634903] env[62558]: DEBUG nova.scheduler.client.report [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1376.139786] env[62558]: DEBUG nova.compute.resource_tracker [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62558) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1376.140209] env[62558]: DEBUG oslo_concurrency.lockutils [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.170s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1376.140253] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1376.140382] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1376.646183] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] There are 11 instances to clean {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1376.646449] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: d11bf400-f80e-4815-9641-6af6bd1b4460] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.149982] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 7c3e82e2-1d08-4985-9b55-de845550c845] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.653664] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: f6d46374-7a19-4079-8257-1ac69ac82b81] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.157567] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 5be1007a-6229-4d45-82d0-53d81a28204c] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.661293] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 301dde86-496e-45b7-a86e-5bda587908db] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.164674] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e26d956-f0f4-4ab1-b23c-fd6859875929] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.668411] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: ac1ec03d-e04b-42eb-a1f3-c14887241bfa] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.172345] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 5082012b-e8ed-41d6-a036-b64cbb1e363c] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.339355] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "9c56d217-086c-4869-baf9-0ec771294c32" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1380.339613] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1380.339780] env[62558]: DEBUG nova.compute.manager [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1380.340694] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961c088d-12c7-46da-a507-5e7f03f5b329 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.347229] env[62558]: DEBUG nova.compute.manager [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62558) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1380.347794] env[62558]: DEBUG nova.objects.instance [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'flavor' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1380.675619] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 5f75bf07-002f-4ed4-930a-fbe6f939cb4f] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.853545] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1380.853820] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-feaad35c-540b-4083-b42e-696e69fc1d70 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.860993] env[62558]: DEBUG oslo_vmware.api [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1380.860993] env[62558]: value = "task-1267476" [ 1380.860993] env[62558]: _type = "Task" [ 1380.860993] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1380.869267] env[62558]: DEBUG oslo_vmware.api [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1381.179031] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e8b3008-693d-4445-937b-8e7db13c7c91] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.370631] env[62558]: DEBUG oslo_vmware.api [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267476, 'name': PowerOffVM_Task, 'duration_secs': 0.202145} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1381.370892] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1381.371090] env[62558]: DEBUG nova.compute.manager [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1381.371864] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6418823-c51c-424a-816f-9132068fa80f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.682623] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] [instance: 4e29a5ef-ef6c-44cf-abaa-6e7ebdbeeb3b] Instance has had 0 of 5 cleanup attempts {{(pid=62558) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.882676] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f6ad78f-9611-443b-a291-abb0bdd2fc21 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1382.185711] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.186178] env[62558]: DEBUG nova.compute.manager [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Cleaning up deleted instances with incomplete migration {{(pid=62558) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1382.688123] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1382.915059] env[62558]: DEBUG nova.objects.instance [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'flavor' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1383.420114] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1383.420524] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1383.420524] env[62558]: DEBUG nova.network.neutron [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1383.420731] env[62558]: DEBUG nova.objects.instance [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'info_cache' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1383.924612] env[62558]: DEBUG nova.objects.base [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Object Instance<9c56d217-086c-4869-baf9-0ec771294c32> lazy-loaded attributes: flavor,info_cache {{(pid=62558) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1384.625545] env[62558]: DEBUG nova.network.neutron [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1385.128214] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1385.631892] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powering on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1385.632264] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab59fcc9-2032-4ad3-bd3d-6d95eda41841 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.639848] env[62558]: DEBUG oslo_vmware.api [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1385.639848] env[62558]: value = "task-1267477" [ 1385.639848] env[62558]: _type = "Task" [ 1385.639848] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.646798] env[62558]: DEBUG oslo_vmware.api [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.150281] env[62558]: DEBUG oslo_vmware.api [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267477, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.650881] env[62558]: DEBUG oslo_vmware.api [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267477, 'name': PowerOnVM_Task, 'duration_secs': 0.573265} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.651298] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powered on the VM {{(pid=62558) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1386.651395] env[62558]: DEBUG nova.compute.manager [None req-e794b0fe-d52a-41b3-b552-5a51de069443 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1386.652123] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103021db-3705-4d9f-b158-7a0cce88175e {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.878948] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f90a46-e0dc-47aa-9186-5d3f14d9457a {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.885549] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Suspending the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1387.885795] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-aaa1c808-644a-431e-a15a-ad7a2713c904 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.891481] env[62558]: DEBUG oslo_vmware.api [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1387.891481] env[62558]: value = "task-1267478" [ 1387.891481] env[62558]: _type = "Task" [ 1387.891481] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1387.899542] env[62558]: DEBUG oslo_vmware.api [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267478, 'name': SuspendVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.186323] env[62558]: DEBUG oslo_service.periodic_task [None req-a861bf76-b7eb-4231-8466-d7a18255dc8a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62558) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1388.402724] env[62558]: DEBUG oslo_vmware.api [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267478, 'name': SuspendVM_Task} progress is 75%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.902487] env[62558]: DEBUG oslo_vmware.api [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267478, 'name': SuspendVM_Task, 'duration_secs': 0.55052} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.903028] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Suspended the VM {{(pid=62558) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1388.903175] env[62558]: DEBUG nova.compute.manager [None req-2c2c3ab1-3c27-425f-923d-7a6e1f6f30ec tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1388.903991] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9b3e17-3613-41eb-9b71-b4361b60ff4d {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.218274] env[62558]: INFO nova.compute.manager [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Resuming [ 1390.218896] env[62558]: DEBUG nova.objects.instance [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'flavor' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.226460] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1391.226876] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquired lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1391.226876] env[62558]: DEBUG nova.network.neutron [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Building network info cache for instance {{(pid=62558) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1391.920611] env[62558]: DEBUG nova.network.neutron [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [{"id": "e858943c-a264-4787-b618-3e9d78a3b6b5", "address": "fa:16:3e:4a:76:f4", "network": {"id": "cff641ea-1726-4bfa-92e0-8c0063668cf6", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-352149384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "688bb982a71a4730aa1652ea837e1b73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape858943c-a2", "ovs_interfaceid": "e858943c-a264-4787-b618-3e9d78a3b6b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1392.423186] env[62558]: DEBUG oslo_concurrency.lockutils [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Releasing lock "refresh_cache-9c56d217-086c-4869-baf9-0ec771294c32" {{(pid=62558) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1392.424225] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04c4c08-da74-414a-a583-70627f24be23 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.431193] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Resuming the VM {{(pid=62558) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1392.431419] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52164fe3-e324-4a36-899b-ea89c35702de {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.437747] env[62558]: DEBUG oslo_vmware.api [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1392.437747] env[62558]: value = "task-1267479" [ 1392.437747] env[62558]: _type = "Task" [ 1392.437747] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.445131] env[62558]: DEBUG oslo_vmware.api [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.948716] env[62558]: DEBUG oslo_vmware.api [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267479, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.448840] env[62558]: DEBUG oslo_vmware.api [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267479, 'name': PowerOnVM_Task, 'duration_secs': 0.674346} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.448840] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Resumed the VM {{(pid=62558) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1393.449252] env[62558]: DEBUG nova.compute.manager [None req-e16d97d7-d1ce-4c44-b795-a778a4741a14 tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Checking state {{(pid=62558) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1393.449742] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f50d56b-b8d3-4fbe-89e8-0970aab012be {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.899054] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "9c56d217-086c-4869-baf9-0ec771294c32" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.899452] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.899586] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "9c56d217-086c-4869-baf9-0ec771294c32-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1394.899771] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1394.899948] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.902231] env[62558]: INFO nova.compute.manager [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Terminating instance [ 1394.904012] env[62558]: DEBUG nova.compute.manager [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Start destroying the instance on the hypervisor. {{(pid=62558) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1394.904218] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Destroying instance {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1394.905059] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abf8944-3edf-4077-a907-584b545f6d2f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.912561] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powering off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1394.912775] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80bf933a-a63e-43ca-98dd-e79ead495772 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.919225] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1394.919225] env[62558]: value = "task-1267480" [ 1394.919225] env[62558]: _type = "Task" [ 1394.919225] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.926262] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.429077] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267480, 'name': PowerOffVM_Task, 'duration_secs': 0.174441} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.429359] env[62558]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Powered off the VM {{(pid=62558) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1395.429533] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Unregistering the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1395.429784] env[62558]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c751895-8270-4416-a657-80dbe8f9a4e6 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.489400] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Unregistered the VM {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1395.489614] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Deleting contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1395.489791] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleting the datastore file [datastore2] 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1395.490111] env[62558]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dffc6ab-c0c6-49bc-a26d-1fe5c3542e5c {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.497107] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for the task: (returnval){ [ 1395.497107] env[62558]: value = "task-1267482" [ 1395.497107] env[62558]: _type = "Task" [ 1395.497107] env[62558]: } to complete. {{(pid=62558) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.504294] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.007283] env[62558]: DEBUG oslo_vmware.api [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Task: {'id': task-1267482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132127} completed successfully. {{(pid=62558) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.007642] env[62558]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted the datastore file {{(pid=62558) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1396.007733] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Deleted contents of the VM from datastore datastore2 {{(pid=62558) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1396.007889] env[62558]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Instance destroyed {{(pid=62558) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1396.008083] env[62558]: INFO nova.compute.manager [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1396.008349] env[62558]: DEBUG oslo.service.loopingcall [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62558) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1396.008549] env[62558]: DEBUG nova.compute.manager [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Deallocating network for instance {{(pid=62558) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1396.008644] env[62558]: DEBUG nova.network.neutron [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] deallocate_for_instance() {{(pid=62558) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1396.439657] env[62558]: DEBUG nova.compute.manager [req-17ba06b1-2644-4b55-98c7-c02589d50e1b req-b151e945-8072-4e5b-bcf5-9ee14e7c1de4 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Received event network-vif-deleted-e858943c-a264-4787-b618-3e9d78a3b6b5 {{(pid=62558) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1396.439866] env[62558]: INFO nova.compute.manager [req-17ba06b1-2644-4b55-98c7-c02589d50e1b req-b151e945-8072-4e5b-bcf5-9ee14e7c1de4 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Neutron deleted interface e858943c-a264-4787-b618-3e9d78a3b6b5; detaching it from the instance and deleting it from the info cache [ 1396.440053] env[62558]: DEBUG nova.network.neutron [req-17ba06b1-2644-4b55-98c7-c02589d50e1b req-b151e945-8072-4e5b-bcf5-9ee14e7c1de4 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1396.920523] env[62558]: DEBUG nova.network.neutron [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Updating instance_info_cache with network_info: [] {{(pid=62558) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1396.942115] env[62558]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7536b225-c7ee-43da-9e22-f3d3c46c5567 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.951710] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adf5e58-1e10-40fa-8d03-c526c00b8437 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.972848] env[62558]: DEBUG nova.compute.manager [req-17ba06b1-2644-4b55-98c7-c02589d50e1b req-b151e945-8072-4e5b-bcf5-9ee14e7c1de4 service nova] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Detach interface failed, port_id=e858943c-a264-4787-b618-3e9d78a3b6b5, reason: Instance 9c56d217-086c-4869-baf9-0ec771294c32 could not be found. {{(pid=62558) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1397.423473] env[62558]: INFO nova.compute.manager [-] [instance: 9c56d217-086c-4869-baf9-0ec771294c32] Took 1.41 seconds to deallocate network for instance. [ 1397.930175] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1397.930494] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1397.930728] env[62558]: DEBUG nova.objects.instance [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lazy-loading 'resources' on Instance uuid 9c56d217-086c-4869-baf9-0ec771294c32 {{(pid=62558) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.526901] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6041673-8a1d-4132-b3fc-3af5250f1dd1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.534211] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791cd282-f600-432d-a768-150c119072d1 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.563159] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d6b6cc-1a2e-4172-aad2-2d36e2c15c1f {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.569628] env[62558]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c286bba1-f763-4756-8872-bb02767f5601 {{(pid=62558) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.581940] env[62558]: DEBUG nova.compute.provider_tree [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed in ProviderTree for provider: dc830c09-1c36-446a-8af3-d3826bec8b3b {{(pid=62558) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1399.084761] env[62558]: DEBUG nova.scheduler.client.report [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Inventory has not changed for provider dc830c09-1c36-446a-8af3-d3826bec8b3b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62558) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1399.590104] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.611350] env[62558]: INFO nova.scheduler.client.report [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Deleted allocations for instance 9c56d217-086c-4869-baf9-0ec771294c32 [ 1400.121751] env[62558]: DEBUG oslo_concurrency.lockutils [None req-0f3776ac-22e8-49b6-a107-28ad0a4c96df tempest-ServerActionsTestJSON-636686932 tempest-ServerActionsTestJSON-636686932-project-member] Lock "9c56d217-086c-4869-baf9-0ec771294c32" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.222s {{(pid=62558) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}